aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorJussi Pakkanen <jpakkane@gmail.com>2017-01-03 13:34:04 -0500
committerGitHub <noreply@github.com>2017-01-03 13:34:04 -0500
commitc025c71ae282cfe33e9fe5d8042e5b511808f2de (patch)
tree6fbc7cd6a2d505540353db55c2a93772b9eef470
parent3b0eec77034868f0461efabcf15a5d418e92c57b (diff)
parentebea1e3bd95c8120dbd72d3629ca242983eef1b8 (diff)
downloadmeson-c025c71ae282cfe33e9fe5d8042e5b511808f2de.zip
meson-c025c71ae282cfe33e9fe5d8042e5b511808f2de.tar.gz
meson-c025c71ae282cfe33e9fe5d8042e5b511808f2de.tar.bz2
Merge pull request #1252 from trhd/master
Some mesontest related improvements.
-rwxr-xr-xmesontest.py254
1 files changed, 134 insertions, 120 deletions
diff --git a/mesontest.py b/mesontest.py
index c7c60c0..af4ae2a 100755
--- a/mesontest.py
+++ b/mesontest.py
@@ -103,9 +103,7 @@ class TestRun():
if self.cmd is None:
res += 'NONE\n'
else:
- res += "\n%s %s\n" % (' '.join(
- ["%s='%s'" % (k, v) for k, v in self.env.items()]),
- ' ' .join(self.cmd))
+ res += "%s%s\n" % (''.join(["%s='%s' " % (k, v) for k, v in self.env.items()]), ' ' .join(self.cmd))
if self.stdo:
res += '--- stdout ---\n'
res += self.stdo
@@ -151,8 +149,10 @@ class TestHarness:
def __init__(self, options):
self.options = options
self.collected_logs = []
- self.failed_tests = []
- self.error_count = 0
+ self.fail_count = 0
+ self.success_count = 0
+ self.skip_count = 0
+ self.timeout_count = 0
self.is_run = False
self.cant_rebuild = False
if self.options.benchmark:
@@ -182,7 +182,6 @@ class TestHarness:
return True
def run_single_test(self, wrap, test):
- failling = False
if test.fname[0].endswith('.jar'):
cmd = ['java', '-jar'] + test.fname
elif not test.is_cross and run_with_mono(test.fname[0]):
@@ -258,21 +257,19 @@ class TestHarness:
stde = decode(stde)
if timed_out:
res = 'TIMEOUT'
- failling = True
+ self.timeout_count += 1
if p.returncode == GNU_SKIP_RETURNCODE:
res = 'SKIP'
- elif (not test.should_fail and p.returncode == 0) or \
- (test.should_fail and p.returncode != 0):
+ self.skip_count += 1
+ elif test.should_fail == bool(p.returncode):
res = 'OK'
+ self.success_count += 1
else:
res = 'FAIL'
- failling = True
+ self.fail_count += 1
returncode = p.returncode
result = TestRun(res, returncode, test.should_fail, duration, stdo, stde, cmd, test.env)
- if failling:
- self.failed_tests.append(result)
-
return result
def print_stats(self, numlen, tests, name, result, i, logfile, jsonlogfile):
@@ -286,20 +283,21 @@ class TestHarness:
result_str += "\n\n" + result.get_log()
if (result.returncode != GNU_SKIP_RETURNCODE) and \
(result.returncode != 0) != result.should_fail:
- self.error_count += 1
if self.options.print_errorlogs:
self.collected_logs.append(result_str)
- logfile.write(result_str)
- write_json_log(jsonlogfile, name, result)
-
- def doit(self):
- if self.is_run:
- raise RuntimeError('Test harness object can only be used once.')
- if not os.path.isfile(self.datafile):
- print('Test data file. Probably this means that you did not run this in the build directory.')
- return 1
- self.is_run = True
- logfilename = self.run_tests(self.options.logbase)
+ if logfile:
+ logfile.write(result_str)
+ if jsonlogfile:
+ write_json_log(jsonlogfile, name, result)
+
+ def print_summary(self, logfile, jsonlogfile):
+ msg = 'Test summary: %d OK, %d FAIL, %d SKIP, %d TIMEOUT' \
+ % (self.success_count, self.fail_count, self.skip_count, self.timeout_count)
+ print(msg)
+ if logfile:
+ logfile.write(msg)
+
+ def print_collected_logs(self):
if len(self.collected_logs) > 0:
if len(self.collected_logs) > 10:
print('\nThe output from 10 first failed tests:\n')
@@ -307,58 +305,100 @@ class TestHarness:
print('\nThe output from the failed tests:\n')
for log in self.collected_logs[:10]:
lines = log.splitlines()
- if len(lines) > 100:
- print(lines[0])
+ if len(lines) > 104:
+ print('\n'.join(lines[0:4]))
print('--- Listing only the last 100 lines from a long log. ---')
- lines = lines[-99:]
+ lines = lines[-100:]
for line in lines:
print(line)
- print('Full log written to %s.' % logfilename)
- return self.error_count
+
+ def doit(self):
+ if self.is_run:
+ raise RuntimeError('Test harness object can only be used once.')
+ if not os.path.isfile(self.datafile):
+ print('Test data file. Probably this means that you did not run this in the build directory.')
+ return 1
+ self.is_run = True
+ tests = self.get_tests()
+ if not tests:
+ return 0
+ self.run_tests(tests)
+ return self.fail_count
def get_tests(self):
with open(self.datafile, 'rb') as f:
tests = pickle.load(f)
+
+ if not tests:
+ print('No tests defined.')
+ return []
+
+ if self.options.suite:
+ tests = [t for t in tests if self.options.suite in t.suite]
+
+ if self.options.args:
+ tests = [t for t in tests if t.name in self.options.args]
+
+ if not tests:
+ print('No suitable tests defined.')
+ return []
+
for test in tests:
test.rebuilt = False
return tests
- def run_tests(self, log_base):
- logfile_base = os.path.join(self.options.wd, 'meson-logs', log_base)
+ def open_log_files(self):
+ if not self.options.logbase or self.options.verbose:
+ return (None, None, None, None)
+
+ logfile_base = os.path.join(self.options.wd, 'meson-logs', self.options.logbase)
+
if self.options.wrapper is None:
- wrap = []
logfilename = logfile_base + '.txt'
jsonlogfilename = logfile_base + '.json'
else:
+ namebase = os.path.split(self.get_wrapper()[0])[1]
+ logfilename = logfile_base + '-' + namebase.replace(' ', '_') + '.txt'
+ jsonlogfilename = logfile_base + '-' + namebase.replace(' ', '_') + '.json'
+
+ jsonlogfile = open(jsonlogfilename, 'w')
+ logfile = open(logfilename, 'w')
+
+ logfile.write('Log of Meson test suite run on %s.\n\n'
+ % datetime.datetime.now().isoformat())
+
+ return (logfile, logfilename, jsonlogfile, jsonlogfilename)
+
+ def get_wrapper(self):
+ wrap = []
+ if self.options.gdb:
+ wrap = ['gdb', '--quiet', '--nh']
+ if self.options.repeat > 1:
+ wrap += ['-ex', 'run', '-ex', 'quit']
+ elif self.options.wrapper:
if isinstance(self.options.wrapper, str):
wrap = self.options.wrapper.split()
else:
wrap = self.options.wrapper
- assert(isinstance(wrap, list))
- namebase = os.path.split(wrap[0])[1]
- logfilename = logfile_base + '-' + namebase.replace(' ', '_') + '.txt'
- jsonlogfilename = logfile_base + '-' + namebase.replace(' ', '_') + '.json'
- tests = self.get_tests()
- if len(tests) == 0:
- print('No tests defined.')
- return
- numlen = len('%d' % len(tests))
- executor = conc.ThreadPoolExecutor(max_workers=self.options.num_processes)
- futures = []
- filtered_tests = filter_tests(self.options.suite, tests)
-
- jsonlogfile = None
- logfile = None
+ assert(isinstance(wrap, list))
+ return wrap
+
+ def get_suites(self, tests):
+ return set([test.suite[0] for test in tests])
+
+ def run_tests(self, tests):
try:
- if not self.options.verbose:
- jsonlogfile = open(jsonlogfilename, 'w')
- logfile = open(logfilename, 'w')
- logfile.write('Log of Meson test suite run on %s.\n\n' %
- datetime.datetime.now().isoformat())
+ executor = None
+ logfile = None
+ jsonlogfile = None
+ futures = []
+ numlen = len('%d' % len(tests))
+ (logfile, logfilename, jsonlogfile, jsonlogfilename) = self.open_log_files()
+ wrap = self.get_wrapper()
for i in range(self.options.repeat):
- for i, test in enumerate(filtered_tests):
+ for i, test in enumerate(tests):
if test.suite[0] == '':
visible_name = test.name
else:
@@ -367,97 +407,66 @@ class TestHarness:
else:
visible_name = test.suite[0] + ' / ' + test.name
- if not test.is_parallel:
- self.drain_futures(futures)
+ if self.options.gdb:
+ test.timeout = None
+ if len(test.cmd_args):
+ wrap.append('--args')
+
+ if not test.is_parallel or self.options.gdb:
+ self.drain_futures(futures, logfile, jsonlogfile)
futures = []
res = self.run_single_test(wrap, test)
- if not self.options.verbose:
- self.print_stats(numlen, filtered_tests, visible_name, res, i,
- logfile, jsonlogfile)
+ self.print_stats(numlen, tests, visible_name, res, i, logfile, jsonlogfile)
else:
+ if not executor:
+ executor = conc.ThreadPoolExecutor(max_workers=self.options.num_processes)
f = executor.submit(self.run_single_test, wrap, test)
- if not self.options.verbose:
- futures.append((f, numlen, filtered_tests, visible_name, i,
- logfile, jsonlogfile))
+ futures.append((f, numlen, tests, visible_name, i, logfile, jsonlogfile))
+ if self.options.repeat > 1 and self.fail_count:
+ break
+ if self.options.repeat > 1 and self.fail_count:
+ break
+
self.drain_futures(futures, logfile, jsonlogfile)
+ self.print_summary(logfile, jsonlogfile)
+ self.print_collected_logs()
+
+ if logfilename:
+ print('Full log written to %s.' % logfilename)
finally:
if jsonlogfile:
jsonlogfile.close()
if logfile:
logfile.close()
- return logfilename
-
-
def drain_futures(self, futures, logfile, jsonlogfile):
for i in futures:
(result, numlen, tests, name, i, logfile, jsonlogfile) = i
- if self.options.repeat > 1 and self.failed_tests:
+ if self.options.repeat > 1 and self.fail_count:
result.cancel()
- elif not self.options.verbose:
- self.print_stats(numlen, tests, name, result.result(), i, logfile, jsonlogfile)
- else:
+ if self.options.verbose:
result.result()
-
- if self.options.repeat > 1 and self.failed_tests:
- if not self.options.verbose:
- for res in self.failed_tests:
- print('Test failed:\n\n-- stdout --\n')
- print(res.stdo)
- print('\n-- stderr --\n')
- print(res.stde)
- return 1
-
- return
+ self.print_stats(numlen, tests, name, result.result(), i, logfile, jsonlogfile)
def run_special(self):
'Tests run by the user, usually something like "under gdb 1000 times".'
if self.is_run:
raise RuntimeError('Can not use run_special after a full run.')
- if self.options.wrapper is not None:
- wrap = self.options.wrapper.split(' ')
- else:
- wrap = []
- if self.options.gdb and len(wrap) > 0:
- print('Can not specify both a wrapper and gdb.')
- return 1
if os.path.isfile('build.ninja'):
subprocess.check_call([environment.detect_ninja(), 'all'])
tests = self.get_tests()
- if self.options.list:
- for i in tests:
- print(i.name)
- return 0
- for t in tests:
- if t.name in self.options.args:
- for i in range(self.options.repeat):
- print('Running: %s %d/%d' % (t.name, i + 1, self.options.repeat))
- if self.options.gdb:
- wrap = ['gdb', '--quiet']
- if len(t.cmd_args) > 0:
- wrap.append('--args')
- if self.options.repeat > 1:
- # The user wants to debug interactively, so no timeout.
- t.timeout = None
- wrap += ['-ex', 'run', '-ex', 'quit']
+ self.run_tests(tests)
+ return self.fail_count
- res = self.run_single_test(wrap, t)
- else:
- res = self.run_single_test(wrap, t)
- if (res.returncode == 0 and res.should_fail) or \
- (res.returncode != 0 and not res.should_fail):
- if not self.options.verbose:
- print('Test failed:\n\n-- stdout --\n')
- print(res.stdo)
- print('\n-- stderr --\n')
- print(res.stde)
- return 1
- return 0
-def filter_tests(suite, tests):
- if suite is None:
- return tests
- return [x for x in tests if suite in x.suite]
+def list_tests(th):
+ tests = th.get_tests()
+ print_suites = True if len(th.get_suites(tests)) != 1 else False
+ for i in tests:
+ if print_suites:
+ print("%s / %s" % (i.suite[0], i.name))
+ else:
+ print("%s" % i.name)
def merge_suite_options(options):
@@ -482,6 +491,7 @@ def merge_suite_options(options):
def run(args):
options = parser.parse_args(args)
+
if options.benchmark:
options.num_processes = 1
@@ -494,12 +504,16 @@ def run(args):
if options.gdb:
options.verbose = True
+ if options.wrapper:
+ print('Must not specify both a wrapper and gdb at the same time.')
+ return 1
options.wd = os.path.abspath(options.wd)
th = TestHarness(options)
if options.list:
- return th.run_special()
+ list_tests(th)
+ return 0
if not options.no_rebuild:
if not th.rebuild_all():
sys.exit(-1)