summaryrefslogtreecommitdiff
path: root/tests/SConscript
diff options
context:
space:
mode:
authorAndreas Sandberg <andreas.sandberg@arm.com>2016-05-31 11:27:00 +0100
committerAndreas Sandberg <andreas.sandberg@arm.com>2016-05-31 11:27:00 +0100
commit44e9b81f7477feec5786de40c60d986e1be74326 (patch)
treea7276aa4cf3d3d6be6f3e4ad491f2d6dc8322b5e /tests/SConscript
parentdbf64aa2c2f803a659ed3cb8a128dbe1f73cfa06 (diff)
downloadgem5-44e9b81f7477feec5786de40c60d986e1be74326.tar.xz
scons: Use the new test framework from scons
Rewrite the SCons script responsible for launching tests to use the new test framework. The biggest visible change after this changeset is that SCons no longer produces a "status" file in test build directories. Instead, it creates a status.pickle file. That file can be inspected using the new tests.py script. Another visible change is that timed out tests are now reported as failed rather than a using a separate summary message. Since the pickle file will remain in the build directory after running tests, it's now possible to convert the test results into other formats. For example: ./tests/tests.py show --format junit -o junit.xml \ `find build/ -name status.pickle` To better facilitate running outside of scons, there is now a set of targets that generate test listings that can be used by the test script. There are currently three targets, quick.list, long.list, and all.list. For example: scons build/ARM/tests/opt/all.list for F in `cat build/ARM/tests/opt/all.list`; do ./tests/tests.py run build/ARM/gem5.opt $F done Change-Id: I2c0f8ca0080d7af737362e198eda4cb3a72e6c36 Signed-off-by: Andreas Sandberg <andreas.sandberg@arm.com> Reviewed-by: Curtis Dunham <curtis.dunham@arm.com>
Diffstat (limited to 'tests/SConscript')
-rw-r--r--tests/SConscript460
1 files changed, 165 insertions, 295 deletions
diff --git a/tests/SConscript b/tests/SConscript
index 9baf0ea75..e47d394f2 100644
--- a/tests/SConscript
+++ b/tests/SConscript
@@ -1,5 +1,17 @@
# -*- mode:python -*-
-
+#
+# Copyright (c) 2016 ARM Limited
+# All rights reserved
+#
+# The license below extends only to copyright in the software and shall
+# not be construed as granting a license to any other intellectual
+# property including but not limited to intellectual property relating
+# to a hardware implementation of the functionality of the software
+# licensed hereunder. You may use the software subject to the license
+# terms below provided that you ensure that this notice is replicated
+# unmodified and in its entirety in all distributions of the software,
+# modified or unmodified, in source code or in binary form.
+#
# Copyright (c) 2004-2006 The Regents of The University of Michigan
# All rights reserved.
#
@@ -28,157 +40,62 @@
#
# Authors: Steve Reinhardt
# Kevin Lim
+# Andreas Sandberg
-import os, signal
-import sys, time
-import glob
from SCons.Script.SConscript import SConsEnvironment
+import os
+import pickle
+import sys
-Import('env')
+sys.path.insert(0, Dir(".").srcnode().abspath)
+import testing.tests as tests
+import testing.results as results
-env['DIFFOUT'] = File('diff-out')
+Import('env')
# get the termcap from the environment
termcap = env['TERMCAP']
# Dict that accumulates lists of tests by category (quick, medium, long)
env.Tests = {}
+gpu_isa = env['TARGET_GPU_ISA'] if env['BUILD_GPU'] else None
+for cat in tests.all_categories:
+ env.Tests[cat] = tuple(
+ tests.get_tests(env["TARGET_ISA"],
+ categories=(cat, ),
+ ruby_protocol=env["PROTOCOL"],
+ gpu_isa=gpu_isa))
-def contents(node):
- return file(str(node)).read()
-
-# functions to parse return value from scons Execute()... not the same
-# as wait() etc., so python built-in os funcs don't work.
-def signaled(status):
- return (status & 0x80) != 0;
-
-def signum(status):
- return (status & 0x7f);
-
-# List of signals that indicate that we should retry the test rather
-# than consider it failed.
-retry_signals = (signal.SIGTERM, signal.SIGKILL, signal.SIGINT,
- signal.SIGQUIT, signal.SIGHUP)
-
-# regular expressions of lines to ignore when diffing outputs
-output_ignore_regexes = (
- '^command line:', # for stdout file
- '^gem5 compiled ', # for stderr file
- '^gem5 started ', # for stderr file
- '^gem5 executing on ', # for stderr file
- '^Simulation complete at', # for stderr file
- '^Listening for', # for stderr file
- 'listening for remote gdb', # for stderr file
- )
-
-output_ignore_args = ' '.join(["-I '"+s+"'" for s in output_ignore_regexes])
-
-output_ignore_args += ' --exclude=stats.txt --exclude=outdiff'
+def color_message(color, msg):
+ return color + msg + termcap.Normal
def run_test(target, source, env):
- """Check output from running test.
+ """Run a test and produce results as a pickle file.
Targets are as follows:
- target[0] : status
+ target[0] : Pickle file
Sources are:
source[0] : gem5 binary
source[1] : tests/run.py script
- source[2] : reference stats file
+ source[2:] : reference files
"""
- # make sure target files are all gone
- for t in target:
- if os.path.exists(t.abspath):
- env.Execute(Delete(t.abspath))
-
tgt_dir = os.path.dirname(str(target[0]))
+ config = tests.ClassicConfig(*tgt_dir.split('/')[-6:])
+ test = tests.ClassicTest(source[0].abspath, tgt_dir, config,
+ timeout=5*60*60,
+ skip_diff_out=True)
+
+ for ref in test.ref_files():
+ out_file = os.path.join(tgt_dir, ref)
+ if os.path.exists(out_file):
+ env.Execute(Delete(out_file))
+
+ with open(target[0].abspath, "wb") as fout:
+ formatter = results.Pickle(fout=fout)
+ formatter.dump_suites([ test.run() ])
- # Base command for running test. We mess around with indirectly
- # referring to files via SOURCES and TARGETS so that scons can mess
- # with paths all it wants to and we still get the right files.
- cmd = '${SOURCES[0]} -d %s -re ${SOURCES[1]} %s' % (tgt_dir, tgt_dir)
-
- # Prefix test run with batch job submission command if appropriate.
- # Batch command also supports timeout arg (in seconds, not minutes).
- timeout = 15 * 60 # used to be a param, probably should be again
- if env['BATCH']:
- cmd = '%s -t %d %s' % (env['BATCH_CMD'], timeout, cmd)
- # The slowest regression (bzip2) requires ~4 hours;
- # 5 hours was chosen to be conservative.
- elif env['TIMEOUT']:
- cmd = 'timeout --foreground 5h %s' % cmd
-
- # Create a default value for the status string, changed as needed
- # based on the status.
- status_str = "passed."
-
- pre_exec_time = time.time()
- status = env.Execute(env.subst(cmd, target=target, source=source))
- if status == 0:
- # gem5 terminated normally.
- # Run diff on output & ref directories to find differences.
- # Exclude the stats file since we will use diff-out on that.
-
- # NFS file systems can be annoying and not have updated yet
- # wait until we see the file modified
- statsdiff = os.path.join(tgt_dir, 'statsdiff')
- m_time = 0
- nap = 0
- while m_time < pre_exec_time and nap < 10:
- try:
- m_time = os.stat(statsdiff).st_mtime
- except OSError:
- pass
- time.sleep(1)
- nap += 1
-
- outdiff = os.path.join(tgt_dir, 'outdiff')
- # tack 'true' on the end so scons doesn't report diff's
- # non-zero exit code as a build error
- diffcmd = 'diff -ubrs %s ${SOURCES[2].dir} %s > %s; true' \
- % (output_ignore_args, tgt_dir, outdiff)
- env.Execute(env.subst(diffcmd, target=target, source=source))
- print "===== Output differences ====="
- print contents(outdiff)
- # Run diff-out on stats.txt file
- diffcmd = '$DIFFOUT ${SOURCES[2]} %s > %s' \
- % (os.path.join(tgt_dir, 'stats.txt'), statsdiff)
- diffcmd = env.subst(diffcmd, target=target, source=source)
- diff_status = env.Execute(diffcmd, strfunction=None)
- # If there is a difference, change the status string to say so
- if diff_status != 0:
- status_str = "CHANGED!"
- print "===== Statistics differences ====="
- print contents(statsdiff)
-
- else: # gem5 exit status != 0
- # Consider it a failed test unless the exit status is 2
- status_str = "FAILED!"
- # gem5 did not terminate properly, so no need to check the output
- if env['TIMEOUT'] and status == 124:
- status_str = "TIMED-OUT!"
- elif signaled(status):
- print 'gem5 terminated with signal', signum(status)
- if signum(status) in retry_signals:
- # Consider the test incomplete; don't create a 'status' output.
- # Hand the return status to scons and let scons decide what
- # to do about it (typically terminate unless run with -k).
- return status
- elif status == 2:
- # The test was skipped, change the status string to say so
- status_str = "skipped."
- else:
- print 'gem5 exited with non-zero status', status
- # complete but failed execution (call to exit() with non-zero
- # status, SIGABORT due to assertion failure, etc.)... fall through
- # and generate FAILED status as if output comparison had failed
-
- # Generate status file contents based on exit status of gem5 and diff-out
- f = file(str(target[0]), 'w')
- print >>f, tgt_dir, status_str
- f.close()
- # done
return 0
def run_test_string(target, source, env):
@@ -188,197 +105,150 @@ def run_test_string(target, source, env):
testAction = env.Action(run_test, run_test_string)
def print_test(target, source, env):
- # print the status with colours to make it easier to see what
- # passed and what failed
- line = contents(source[0])
-
- # split the line to words and get the last one
- words = line.split()
- status = words[-1]
-
- # if the test failed make it red, if it passed make it green, and
- # skip the punctuation
- if status == "FAILED!" or status == "TIMED-OUT!":
- status = termcap.Red + status[:-1] + termcap.Normal + status[-1]
- elif status == "CHANGED!":
- status = termcap.Yellow + status[:-1] + termcap.Normal + status[-1]
- elif status == "passed.":
- status = termcap.Green + status[:-1] + termcap.Normal + status[-1]
- elif status == "skipped.":
- status = termcap.Cyan + status[:-1] + termcap.Normal + status[-1]
-
- # put it back in the list and join with space
- words[-1] = status
- line = " ".join(words)
-
- print '***** ' + line
+ """Run a test and produce results as a pickle file.
+
+ Targets are as follows:
+ target[*] : Dummy targets
+
+ Sources are:
+ source[0] : Pickle file
+
+ """
+ with open(source[0].abspath, "rb") as fin:
+ result = pickle.load(fin)
+
+ assert len(result) == 1
+ result = result[0]
+
+ run = result.results[0]
+ assert run.name == "gem5"
+
+ formatter = None
+ if not run:
+ status = color_message(termcap.Red, "FAILED!")
+ formatter = results.Text()
+ elif run.skipped():
+ status = color_message(termcap.Cyan, "skipped.")
+ elif result:
+ status = color_message(termcap.Green, "passed.")
+ else:
+ status = color_message(termcap.Yellow, "CHANGED!")
+ formatter = results.Text()
+
+ if formatter:
+ formatter.dump_suites([result])
+
+ print "***** %s: %s" % (source[0].dir, status)
return 0
-printAction = env.Action(print_test, strfunction = None)
-
-# Static vars for update_test:
-# - long-winded message about ignored sources
-ignore_msg = '''
-Note: The following file(s) will not be copied. New non-standard
- output files must be copied manually once before --update-ref will
- recognize them as outputs. Otherwise they are assumed to be
- inputs and are ignored.
-'''
-# - reference files always needed
-needed_files = set(['simout', 'simerr', 'stats.txt', 'config.ini'])
-# - source files we always want to ignore
-known_ignores = set(['status', 'outdiff', 'statsdiff'])
+printAction = env.Action(print_test, strfunction=None)
def update_test(target, source, env):
- """Update reference test outputs.
+ """Update test reference data
- Target is phony. First two sources are the ref & new stats.txt file
- files, respectively. We actually copy everything in the
- respective directories except the status & diff output files.
+ Targets are as follows:
+ target[0] : Dummy file
+ Sources are:
+ source[0] : Pickle file
"""
- dest_dir = str(source[0].get_dir())
- src_dir = str(source[1].get_dir())
- dest_files = set(os.listdir(dest_dir))
- src_files = set(os.listdir(src_dir))
- # Copy all of the required files plus any existing dest files.
- wanted_files = needed_files | dest_files
- missing_files = wanted_files - src_files
- if len(missing_files) > 0:
- print " WARNING: the following file(s) are missing " \
- "and will not be updated:"
- print " ", " ,".join(missing_files)
- copy_files = wanted_files - missing_files
- warn_ignored_files = (src_files - copy_files) - known_ignores
- if len(warn_ignored_files) > 0:
- print ignore_msg,
- print " ", ", ".join(warn_ignored_files)
- for f in copy_files:
- if f in dest_files:
- print " Replacing file", f
- dest_files.remove(f)
- else:
- print " Creating new file", f
- copyAction = Copy(os.path.join(dest_dir, f), os.path.join(src_dir, f))
- copyAction.strfunction = None
- env.Execute(copyAction)
+
+ src_dir = os.path.dirname(str(source[0]))
+ config = tests.ClassicConfig(*src_dir.split('/')[-6:])
+ test = tests.ClassicTest(source[0].abspath, src_dir, config)
+ ref_dir = test.ref_dir
+
+ with open(source[0].abspath, "rb") as fin:
+ result = pickle.load(fin)
+
+ assert len(result) == 1
+ result = result[0]
+
+ run = result.results[0]
+ assert run.name == "gem5"
+
+ if run.skipped():
+ print "*** %s: %s: Test skipped, not updating." % (
+ source[0].dir, color_message(termcap.Yellow, "WARNING"), )
+ return 0
+ elif result:
+ print "*** %s: %s: Test successful, not updating." % (
+ source[0].dir, color_message(termcap.Green, "skipped"), )
+ return 0
+ elif not run.success():
+ print "*** %s: %s: Test failed, not updating." % (
+ source[0].dir, color_message(termcap.Red, "ERROR"), )
+ return 1
+
+ print "** Updating %s" % (test, )
+ test.update_ref()
+
return 0
def update_test_string(target, source, env):
- return env.subst("Updating ${SOURCES[0].dir} from ${SOURCES[1].dir}",
+ return env.subst("Updating ${SOURCES[0].dir}",
target=target, source=source)
updateAction = env.Action(update_test, update_test_string)
-def test_builder(env, ref_dir):
+def test_builder(test_tuple):
"""Define a test."""
- path = list(ref_dir.split('/'))
+ out_dir = "/".join(test_tuple)
+ binary = env.M5Binary.abspath
+ test = tests.ClassicTest(binary, out_dir, test_tuple)
+
+ def tgt(name):
+ return os.path.join(out_dir, name)
- # target path (where test output goes) consists of category, mode,
- # name, isa, opsys, and config (skips the 'ref' component)
- assert(path.pop(-4) == 'ref')
- tgt_dir = os.path.join(*path[-6:])
+ def ref(name):
+ return os.path.join(test.ref_dir, name)
- # local closure for prepending target path to filename
- def tgt(f):
- return os.path.join(tgt_dir, f)
+ pickle_file = tgt("status.pickle")
+ targets = [
+ pickle_file,
+ ]
- ref_stats = os.path.join(ref_dir, 'stats.txt')
- new_stats = tgt('stats.txt')
- status_file = tgt('status')
+ sources = [
+ env.M5Binary,
+ "run.py",
+ ] + [ ref(f) for f in test.ref_files() ]
- env.Command([status_file, new_stats],
- [env.M5Binary, 'run.py', ref_stats],
- testAction)
+ env.Command(targets, sources, testAction)
# phony target to echo status
if GetOption('update_ref'):
- p = env.Command(tgt('_update'),
- [ref_stats, new_stats, status_file],
- updateAction)
+ p = env.Command(tgt("_update"), [pickle_file], updateAction)
else:
- p = env.Command(tgt('_print'), [status_file], printAction)
+ p = env.Command(tgt("_print"), [pickle_file], printAction)
env.AlwaysBuild(p)
+def list_tests(target, source, env):
+ """Create a list of tests
-# Figure out applicable configs based on build type
-configs = []
-if env['TARGET_ISA'] == 'alpha':
- configs += ['tsunami-simple-atomic',
- 'tsunami-simple-timing',
- 'tsunami-simple-atomic-dual',
- 'tsunami-simple-timing-dual',
- 'twosys-tsunami-simple-atomic',
- 'tsunami-o3', 'tsunami-o3-dual',
- 'tsunami-minor', 'tsunami-minor-dual',
- 'tsunami-switcheroo-full']
-if env['TARGET_ISA'] == 'sparc':
- configs += ['t1000-simple-atomic',
- 't1000-simple-timing']
-if env['TARGET_ISA'] == 'arm':
- configs += ['simple-atomic-dummychecker',
- 'o3-timing-checker',
- 'realview-simple-atomic',
- 'realview-simple-atomic-dual',
- 'realview-simple-atomic-checkpoint',
- 'realview-simple-timing',
- 'realview-simple-timing-dual',
- 'realview-o3',
- 'realview-o3-checker',
- 'realview-o3-dual',
- 'realview-minor',
- 'realview-minor-dual',
- 'realview-switcheroo-atomic',
- 'realview-switcheroo-timing',
- 'realview-switcheroo-o3',
- 'realview-switcheroo-full',
- 'realview64-simple-atomic',
- 'realview64-simple-atomic-checkpoint',
- 'realview64-simple-atomic-dual',
- 'realview64-simple-timing',
- 'realview64-simple-timing-dual',
- 'realview64-o3',
- 'realview64-o3-checker',
- 'realview64-o3-dual',
- 'realview64-minor',
- 'realview64-minor-dual',
- 'realview64-switcheroo-atomic',
- 'realview64-switcheroo-timing',
- 'realview64-switcheroo-o3',
- 'realview64-switcheroo-full']
-if env['TARGET_ISA'] == 'x86' and not env['BUILD_GPU']:
- configs += ['pc-simple-atomic',
- 'pc-simple-timing',
- 'pc-o3-timing',
- 'pc-switcheroo-full']
-
-if env['TARGET_ISA'] == 'x86' and env['BUILD_GPU'] and \
- env['TARGET_GPU_ISA'] == 'hsail':
- configs += ['gpu']
- if env['PROTOCOL'] == 'GPU_RfO':
- configs += ['gpu-randomtest']
-else:
- configs += ['simple-atomic', 'simple-atomic-mp',
- 'simple-timing', 'simple-timing-mp',
- 'minor-timing', 'minor-timing-mp',
- 'o3-timing', 'o3-timing-mt', 'o3-timing-mp',
- 'rubytest', 'memcheck', 'memtest', 'memtest-filter',
- 'tgen-simple-mem', 'tgen-dram-ctrl']
-
- configs += ['learning-gem5-p1-simple', 'learning-gem5-p1-two-level']
-
-if env['PROTOCOL'] != 'None':
- if env['PROTOCOL'] == 'MI_example':
- configs += [c + "-ruby" for c in configs]
- else:
- configs = [c + "-ruby-" + env['PROTOCOL'] for c in configs]
-
-src = Dir('.').srcdir
-for config in configs:
- dirs = src.glob('*/*/*/ref/%s/*/%s' % (env['TARGET_ISA'], config))
- for d in dirs:
- d = str(d)
- if not os.path.exists(os.path.join(d, 'skip')):
- test_builder(env, d)
+ Targets are as follows:
+ target[0] : List file (e.g., tests/opt/all.list, tests/opt/quick.list)
+
+ Sources are: -
+
+ """
+
+ tgt_name = os.path.basename(str(target[0]))
+ base, ext = os.path.splitext(tgt_name)
+ categories = tests.all_categories if base == "all" else (base, )
+
+ with open(target[0].abspath, "w") as fout:
+ for cat in categories:
+ for test in env.Tests[cat]:
+ print >> fout,"/".join(test)
+
+ return 0
+
+testListAction = env.Action(list_tests, strfunction=None)
+
+env.Command("all.list", tuple(), testListAction)
+for cat, test_list in env.Tests.items():
+ env.Command("%s.list" % cat, tuple(), testListAction)
+ for test in test_list:
+ test_builder(test)