From a6e7118c05fa19c5afa92dd0f96524047e886b90 Mon Sep 17 00:00:00 2001 From: Erhan Kurubas Date: Wed, 4 Jun 2025 23:12:20 +0200 Subject: [PATCH] feat(sysview): Add SEGGER multicore trace file support --- tools/esp_app_trace/sysviewtrace_proc.py | 188 ++++++++++++++++++++--- 1 file changed, 164 insertions(+), 24 deletions(-) diff --git a/tools/esp_app_trace/sysviewtrace_proc.py b/tools/esp_app_trace/sysviewtrace_proc.py index d442dfc233..61ce273d70 100755 --- a/tools/esp_app_trace/sysviewtrace_proc.py +++ b/tools/esp_app_trace/sysviewtrace_proc.py @@ -1,6 +1,6 @@ #!/usr/bin/env python # -# SPDX-FileCopyrightText: 2019-2022 Espressif Systems (Shanghai) CO LTD +# SPDX-FileCopyrightText: 2019-2025 Espressif Systems (Shanghai) CO LTD # SPDX-License-Identifier: Apache-2.0 # # This is python script to process various types trace data streams in SystemView format. @@ -14,34 +14,139 @@ import logging import os.path import signal import sys +import tempfile import traceback import espytrace.apptrace as apptrace import espytrace.sysview as sysview -def main(): +def is_segger_multicore_format(file_path): + """Check if the file has offsets in header""" + try: + with open(file_path, 'rb') as f: + header = f.read(200) + header_str = header.decode('utf-8', errors='ignore') - verbosity_levels = [ - logging.CRITICAL, - logging.ERROR, - logging.WARNING, - logging.INFO, - logging.DEBUG - ] + if ( + '; Version SEGGER SystemViewer' in header_str + and '; Author Espressif Inc' in header_str + and '; Offset Core0' in header_str + and '; Offset Core1' in header_str + ): + logging.info('Detected SEGGER multicore format in file:', file_path) + return True + except Exception as e: + logging.error('Error checking file format:', e) + return False + + +def split_segger_multicore_file(file_path): + """Split SEGGER multicore file into separate core files.""" + try: + with open(file_path, 'rb') as f: + # Read first few lines to get offsets for each core + header = f.read(200) + header_str = header.decode('utf-8', errors='ignore') + + core0_offset = None + core1_offset = None + for line in header_str.split('\n'): + if '; Offset Core0' in line: + core0_offset = int(line.strip().split()[-1]) + elif '; Offset Core1' in line: + core1_offset = int(line.strip().split()[-1]) + + if core0_offset is None or core1_offset is None: + logging.error('Failed to parse core offsets') + return None, None + + # Read the entire file + f.seek(0) + data = f.read() + + # Find first 10 sync bytes start offset + sync_start = data.find(b'\x00' * 10) + if sync_start == -1: + logging.error('Sync bytes not found') + return None, None + + core0_offset = sync_start + core1_offset += sync_start + + # Parse original header and get version from there, if not found, use default version + version = 'V3.60' + for line in header_str.split('\n'): + if '; Version SEGGER SystemViewer' in line: + version = line.strip().split()[-1] + break + + # Rebuild header without offset lines + core_header = f';\n; Version SEGGER SystemViewer {version}\n; Author Espressif Inc\n;\n' + core_header = core_header.encode('utf-8') + core_base = core_header + b'\x00' * 10 + + core0_data = core_base + data[core0_offset:core1_offset] + core1_data = core_base + data[core1_offset:] + + core0_file = tempfile.NamedTemporaryFile(delete=False, suffix='.svdat') + core1_file = tempfile.NamedTemporaryFile(delete=False, suffix='.svdat') + + core0_file.write(core0_data) + core1_file.write(core1_data) + + core0_file.close() + core1_file.close() + + return core0_file.name, core1_file.name + + except Exception as e: + logging.error('Failed to process files:', e) + return None, None + + +def main(): + verbosity_levels = [logging.CRITICAL, logging.ERROR, logging.WARNING, logging.INFO, logging.DEBUG] parser = argparse.ArgumentParser(description='ESP32 SEGGER SystemView Trace Parsing Tool') - parser.add_argument('trace_sources', help='Trace data sources. Format: [file://]/path/to/file.', nargs='+', type=str) + parser.add_argument( + 'trace_sources', help='Trace data sources. Format: [file://]/path/to/file.', nargs='+', type=str + ) parser.add_argument('--elf-file', '-b', help='Path to program ELF file.', type=str, default='') parser.add_argument('--tmo', '-w', help='Data wait timeout in sec. -1: infinite, 0: no wait', type=int, default=0) parser.add_argument('--dump-events', '-d', help='Dump all events.', action='store_true') - parser.add_argument('--print-events', '-p', help='Print events of selected types. By default only reports are printed', action='store_true') - parser.add_argument('--include-events', '-i', help='Events types to be included into report.', type=str, choices=['heap', 'log', 'all'], default='all') + parser.add_argument( + '--print-events', + '-p', + help='Print events of selected types. By default only reports are printed', + action='store_true', + ) + parser.add_argument( + '--include-events', + '-i', + help='Events types to be included into report.', + type=str, + choices=['heap', 'log', 'all'], + default='all', + ) parser.add_argument('--toolchain', '-t', help='Toolchain prefix.', type=str, default='xtensa-esp32-elf-') - parser.add_argument('--events-map', '-e', help='Events map file.', type=str, default=os.path.join(os.path.dirname(__file__), 'SYSVIEW_FreeRTOS.txt')) + parser.add_argument( + '--events-map', + '-e', + help='Events map file.', + type=str, + default=os.path.join(os.path.dirname(__file__), 'SYSVIEW_FreeRTOS.txt'), + ) parser.add_argument('--to-json', '-j', help='Print JSON.', action='store_true', default=False) - parser.add_argument('--verbose', '-v', help='Verbosity level. Default 1', choices=range(0, len(verbosity_levels)), type=int, default=1) + parser.add_argument( + '--verbose', + '-v', + help='Verbosity level. Default 1', + choices=range(0, len(verbosity_levels)), + type=int, + default=1, + ) args = parser.parse_args() def sig_int_handler(signum, frame): @@ -60,17 +165,33 @@ def main(): logging.basicConfig(level=verbosity_levels[args.verbose], format='[%(levelname)s] %(message)s') + segger_files = [] + # Only check for SEGGER format if there's exactly one trace source + if len(args.trace_sources) == 1: + trace_source = args.trace_sources[0] + if is_segger_multicore_format(trace_source): + core0_file, core1_file = split_segger_multicore_file(trace_source) + if core0_file and core1_file: + segger_files.extend([core0_file, core1_file]) + args.trace_sources = segger_files + else: + sys.exit(2) + # parse trace files parsers = [] for i, trace_source in enumerate(args.trace_sources): try: parser = sysview.SysViewMultiTraceDataParser(print_events=False, core_id=i) if include_events['heap']: - parser.add_stream_parser(sysview.SysViewTraceDataParser.STREAMID_HEAP, - sysview.SysViewHeapTraceDataParser(print_events=False, core_id=i)) + parser.add_stream_parser( + sysview.SysViewTraceDataParser.STREAMID_HEAP, + sysview.SysViewHeapTraceDataParser(print_events=False, core_id=i), + ) if include_events['log']: - parser.add_stream_parser(sysview.SysViewTraceDataParser.STREAMID_LOG, - sysview.SysViewLogTraceDataParser(print_events=False, core_id=i)) + parser.add_stream_parser( + sysview.SysViewTraceDataParser.STREAMID_LOG, + sysview.SysViewLogTraceDataParser(print_events=False, core_id=i), + ) parsers.append(parser) except Exception as e: logging.error('Failed to create data parser (%s)!', e) @@ -97,13 +218,21 @@ def main(): # merge and process traces try: - proc = sysview.SysViewMultiStreamTraceDataProcessor(traces=parsers, print_events=args.dump_events, keep_all_events=True if args.to_json else False) + proc = sysview.SysViewMultiStreamTraceDataProcessor( + traces=parsers, print_events=args.dump_events, keep_all_events=True if args.to_json else False + ) if include_events['heap']: - proc.add_stream_processor(sysview.SysViewTraceDataParser.STREAMID_HEAP, - sysview.SysViewHeapTraceDataProcessor(args.toolchain, args.elf_file, root_proc=proc, print_heap_events=args.print_events)) + proc.add_stream_processor( + sysview.SysViewTraceDataParser.STREAMID_HEAP, + sysview.SysViewHeapTraceDataProcessor( + args.toolchain, args.elf_file, root_proc=proc, print_heap_events=args.print_events + ), + ) if include_events['log']: - proc.add_stream_processor(sysview.SysViewTraceDataParser.STREAMID_LOG, - sysview.SysViewLogTraceDataProcessor(root_proc=proc, print_log_events=args.print_events)) + proc.add_stream_processor( + sysview.SysViewTraceDataParser.STREAMID_LOG, + sysview.SysViewLogTraceDataProcessor(root_proc=proc, print_log_events=args.print_events), + ) except Exception as e: logging.error('Failed to create data processor (%s)!', e) traceback.print_exc() @@ -119,11 +248,22 @@ def main(): sys.exit(2) finally: if args.to_json: - print(json.dumps(proc, cls=sysview.SysViewTraceDataJsonEncoder, indent=4, separators=(',', ': '), sort_keys=True)) + print( + json.dumps( + proc, cls=sysview.SysViewTraceDataJsonEncoder, indent=4, separators=(',', ': '), sort_keys=True + ) + ) else: proc.print_report() proc.cleanup() + if segger_files: + for file in segger_files: + try: + os.remove(file) + except Exception as e: + logging.warning('Failed to remove temporary file %s: %s', file, e) + if __name__ == '__main__': main()