[LNT] r235743 - Revert "Make the score sample type work with reruns."
Chris Matthews
cmatthews5 at apple.com
Fri Apr 24 11:47:37 PDT 2015
Author: cmatthews
Date: Fri Apr 24 13:47:36 2015
New Revision: 235743
URL: http://llvm.org/viewvc/llvm-project?rev=235743&view=rev
Log:
Revert "Make the score sample type work with reruns."
I accidently combined a whitespace cleanup with this patch.
This reverts commit r235659.
Modified:
lnt/trunk/lnt/tests/nt.py
Modified: lnt/trunk/lnt/tests/nt.py
URL: http://llvm.org/viewvc/llvm-project/lnt/trunk/lnt/tests/nt.py?rev=235743&r1=235742&r2=235743&view=diff
==============================================================================
--- lnt/trunk/lnt/tests/nt.py (original)
+++ lnt/trunk/lnt/tests/nt.py Fri Apr 24 13:47:36 2015
@@ -77,7 +77,7 @@ class TestConfiguration(object):
self.opts = opts
self.__dict__.update(opts)
self.start_time = start_time
-
+
# Report directory cache.
self._report_dir = None
# Compiler interrogation is a lot of work, this will cache it.
@@ -93,7 +93,7 @@ class TestConfiguration(object):
is either "build" or a timestamped directory based on """
if self._report_dir is not None:
return self._report_dir
-
+
if self.timestamp_build:
ts = self.start_time.replace(' ','_').replace(':','-')
build_dir_name = "test-%s" % ts
@@ -397,7 +397,7 @@ class TestConfiguration(object):
print "make", name, value
make_variables[name] = value
-
+
# Set remote execution variables, if used.
if self.remote:
# make a copy of args for report, without remote options.
@@ -419,7 +419,7 @@ class TestConfiguration(object):
make_variables['USE_PERF'] = '1'
return make_variables, public_vars
-
+
###
def scan_for_test_modules(config):
@@ -610,7 +610,7 @@ def execute_nt_tests(test_log, make_vari
(allow default)
(debug deny)
-;; Deny all file writes by default.
+;; Deny all file writes by default.
(deny file-write*)
;; Deny all network access by default.
@@ -813,7 +813,7 @@ def prepare_build_dir(config, iteration)
else:
needs_clean = False
os.mkdir(build_dir)
-
+
# Unless not using timestamps, we require the basedir not to exist.
if needs_clean and config.timestamp_build:
fatal('refusing to reuse pre-existing build dir %r' % build_dir)
@@ -825,14 +825,14 @@ def update_tools(make_variables, config,
print >>sys.stderr, '%s: building test-suite tools' % (timestamp(),)
args = ['make', 'tools']
args.extend('%s=%s' % (k,v) for k,v in make_variables.items())
- build_tools_log_path = os.path.join(config.build_dir(iteration),
+ build_tools_log_path = os.path.join(config.build_dir(iteration),
'build-tools.log')
build_tools_log = open(build_tools_log_path, 'w')
print >>build_tools_log, '%s: running: %s' % (timestamp(),
' '.join('"%s"' % a
for a in args))
build_tools_log.flush()
- res = execute_command(build_tools_log, config.build_dir(iteration),
+ res = execute_command(build_tools_log, config.build_dir(iteration),
args, config.report_dir)
build_tools_log.close()
if res != 0:
@@ -885,7 +885,7 @@ def copy_missing_makefiles(config, based
def run_test(nick_prefix, iteration, config):
print >>sys.stderr, "%s: checking source versions" % (
timestamp(),)
-
+
test_suite_source_version = get_source_version(config.test_suite_root)
# Compute the make variables.
@@ -929,10 +929,10 @@ def run_test(nick_prefix, iteration, con
if config.only_test is not None and not config.only_test.startswith("LNTBased"):
copy_missing_makefiles(config, basedir)
- # If running without LLVM, make sure tools are up to date.
+ # If running without LLVM, make sure tools are up to date.
if config.without_llvm:
update_tools(make_variables, config, iteration)
-
+
# Always blow away any existing report.
build_report_path = config.build_report_path(iteration)
if os.path.exists(build_report_path):
@@ -1026,7 +1026,7 @@ def run_test(nick_prefix, iteration, con
machdep_info = machine_info
else:
machdep_info = run_info
-
+
machdep_info['uname'] = capture(["uname","-a"], include_stderr=True).strip()
machdep_info['name'] = capture(["uname","-n"], include_stderr=True).strip()
@@ -1178,7 +1178,7 @@ def _execute_test_again(config, test_nam
"TEST." + config.test_style + ".report")
result_path = os.path.join(config.build_dir(None),
test_path, "Output",
- test_name + "." + config.test_style + ".report.csv")
+ test_name + "." + config.test_style + ".report.csv")
gen_report_template = "{gen} -csv {schema} < {input} > {output}"
gen_cmd = gen_report_template.format(gen=config.generate_report_script,
@@ -1215,7 +1215,6 @@ LOCAL_EXEC_STATUS = "exec.status"
# Server results have both status and performance in each entry
SERVER_COMPILE_RESULT = "compile_time"
SERVER_EXEC_RESULT = "execution_time"
-SERVER_SCORE_RESULT = "score"
class PastRunData(object):
@@ -1274,14 +1273,14 @@ class PastRunData(object):
def _process_reruns(config, server_reply, local_results):
"""Rerun each benchmark which the server reported "changed", N more
- times.
+ times.
"""
try:
server_results = server_reply['test_results'][0]['results']
except KeyError:
# Server might send us back an error.
if server_reply.get('error', None):
- warning("Server returned an error:" +
+ warning("Server returned an error:" +
server_reply['error'])
fatal("No Server results. Cannot do reruns.")
logging.fatal()
@@ -1303,7 +1302,7 @@ def _process_reruns(config, server_reply
updating_entry = collated_results.get(test_name,
PastRunData(test_name))
-
+
# Filter out "LNTBased" benchmarks for rerun, they
# won't work. LNTbased look like nts.module.test
# So just filter out anything with .
@@ -1326,7 +1325,7 @@ def _process_reruns(config, server_reply
# Now add on top the server results to any entry we already have.
for full_name, results_status, perf_status in server_results:
fields = full_name.split(".")
- test_name = '.'.join(fields[:-1])
+ test_name = '.'.join(fields[:-1])
test_type = fields[-1]
new_entry = collated_results.get(test_name, None)
@@ -1338,7 +1337,7 @@ def _process_reruns(config, server_reply
if SERVER_COMPILE_RESULT in test_type:
if new_entry.compile_status is None:
new_entry.compile_status = results_status
- elif SERVER_EXEC_RESULT in test_type or SERVER_SCORE_RESULT in test_type:
+ elif SERVER_EXEC_RESULT in test_type:
if new_entry.execution_status is None:
# If the server has not seen the test before, it will return
# None for the performance results analysis. In this case we
@@ -1349,7 +1348,7 @@ def _process_reruns(config, server_reply
derived_perf_status = perf_status
new_entry.execution_status = derived_perf_status
else:
- assert False, "Unexpected server result type." + test_type
+ assert False, "Unexpected server result type."
collated_results[test_name] = new_entry
# Double check that all values are there for all tests.
@@ -1730,7 +1729,7 @@ class NTTest(builtintest.BuiltinTest):
if opts.cxx_under_test is not None:
opts.cxx_under_test = resolve_command_path(opts.cxx_under_test)
-
+
# Always set cxx_under_test, since it may be used as the linker even
# when not testing C++ code.
if opts.cxx_under_test is None:
@@ -1786,7 +1785,7 @@ class NTTest(builtintest.BuiltinTest):
parser.error('--remote is required with --remote-port')
if opts.remote_user is not None:
parser.error('--remote is required with --remote-user')
-
+
# libLTO should exist, if given.
if opts.liblto_path:
if not os.path.exists(opts.liblto_path):
More information about the llvm-commits
mailing list