1#!/usr/bin/env python3 2# Copyright (C) 2018 The Android Open Source Project 3# 4# Licensed under the Apache License, Version 2.0 (the "License"); 5# you may not use this file except in compliance with the License. 6# You may obtain a copy of the License at 7# 8# http://www.apache.org/licenses/LICENSE-2.0 9# 10# Unless required by applicable law or agreed to in writing, software 11# distributed under the License is distributed on an "AS IS" BASIS, 12# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13# See the License for the specific language governing permissions and 14# limitations under the License. 15 16from __future__ import absolute_import 17from __future__ import division 18from __future__ import print_function 19 20import argparse 21import datetime 22import json 23import os 24import re 25import signal 26import sys 27 28ROOT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) 29sys.path.append(os.path.join(ROOT_DIR)) 30 31from python.generators.diff_tests.testing import TestType 32from python.generators.diff_tests.utils import ctrl_c_handler 33from python.generators.diff_tests.runner import DiffTestsRunner 34 35 36def main(): 37 signal.signal(signal.SIGINT, ctrl_c_handler) 38 parser = argparse.ArgumentParser() 39 parser.add_argument('--test-type', type=str, default='all') 40 parser.add_argument('--trace-descriptor', type=str) 41 parser.add_argument('--metrics-descriptor', nargs='+', type=str) 42 parser.add_argument('--chrome-track-event-descriptor', type=str, default=None) 43 parser.add_argument('--test-extensions', type=str, default=None) 44 parser.add_argument('--winscope-extensions', type=str, default=None) 45 parser.add_argument('--perf-file', type=str) 46 parser.add_argument( 47 '--override-sql-module', type=str, action='append', default=[]) 48 parser.add_argument('--test-dir', type=str, default=ROOT_DIR) 49 parser.add_argument( 50 '--name-filter', 51 default='.*', 52 type=str, 53 help='Filter the name of the tests to run (regex syntax)') 54 parser.add_argument( 55 '--keep-input', 56 action='store_true', 57 help='Save the (generated) input pb file for debugging') 58 parser.add_argument( 59 '--rebase', 60 action='store_true', 61 help='Update the expected output file with the actual result') 62 parser.add_argument( 63 '--quiet', action='store_true', help='Only print if the test failed.') 64 parser.add_argument( 65 '--no-colors', action='store_true', help='Print without coloring') 66 parser.add_argument( 67 'trace_processor', type=str, help='location of trace processor binary') 68 args = parser.parse_args() 69 70 out_path = os.path.dirname(args.trace_processor) 71 if args.chrome_track_event_descriptor is None: 72 args.chrome_track_event_descriptor = os.path.join( 73 out_path, 'gen', 'protos', 'third_party', 'chromium', 74 'chrome_track_event.descriptor') 75 if args.test_extensions is None: 76 args.test_extensions = os.path.join(out_path, 'gen', 'protos', 'perfetto', 77 'trace', 'test_extensions.descriptor') 78 if args.winscope_extensions is None: 79 args.winscope_extensions = os.path.join(out_path, 'gen', 'protos', 80 'perfetto', 'trace', 'android', 81 'winscope.descriptor') 82 83 test_runner = DiffTestsRunner(args.name_filter, args.trace_processor, 84 args.trace_descriptor, args.no_colors, 85 args.override_sql_module, args.test_dir, 86 args.quiet) 87 sys.stderr.write(f"[==========] Running {len(test_runner.tests)} tests.\n") 88 89 results = test_runner.run_all_tests(args.metrics_descriptor, 90 args.chrome_track_event_descriptor, 91 args.test_extensions, 92 args.winscope_extensions, args.keep_input, 93 args.rebase) 94 sys.stderr.write(results.str(args.no_colors, len(test_runner.tests))) 95 96 if args.rebase: 97 sys.stderr.write(results.rebase_str()) 98 99 if len(results.test_failures) > 0: 100 return 1 101 102 if args.perf_file: 103 test_dir = os.path.join(ROOT_DIR, 'test') 104 trace_processor_dir = os.path.join(test_dir, 'trace_processor') 105 106 metrics = [] 107 sorted_data = sorted( 108 results.perf_data, key=lambda x: (x.test.type.name, x.test.name)) 109 for perf_args in sorted_data: 110 metrics.append({ 111 'metric': 'tp_perf_test_ingest_time', 112 'value': float(perf_args.ingest_time_ns) / 1.0e9, 113 'unit': 's', 114 'tags': { 115 'test_name': perf_args.test.name, 116 'test_type': perf_args.test.type.name, 117 }, 118 'labels': {}, 119 }) 120 metrics.append({ 121 'metric': 'perf_test_real_time', 122 'value': float(perf_args.real_time_ns) / 1.0e9, 123 'unit': 's', 124 'tags': { 125 'test_name': perf_args.test.name, 126 'test_type': perf_args.test.type.name, 127 }, 128 'labels': {}, 129 }) 130 131 output_data = {'metrics': metrics} 132 with open(args.perf_file, 'w+') as perf_file: 133 perf_file.write(json.dumps(output_data, indent=2)) 134 return 0 135 136 137if __name__ == '__main__': 138 sys.exit(main()) 139