-
Notifications
You must be signed in to change notification settings - Fork 7
/
Copy pathcorrectness-test.py
executable file
·123 lines (97 loc) · 4.91 KB
/
correctness-test.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
#!/usr/bin/env python
# MHS algorithm correctness tester
# Copyright Vera-Licona Research Group (C) 2015
# Author: Andrew Gainer-Dewar, Ph.D. <[email protected]>
# This file is part of MHSGenerationAlgorithms.
# MHSGenerationAlgorithms is free software: you can redistribute it
# and/or modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation, either version 3 of
# the License, or (at your option) any later version.
#
# MHSGenerationAlgorithms is distributed in the hope that it will be
# useful, but WITHOUT ANY WARRANTY; without even the implied warranty
# of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
import argparse
import json
import pyalgorun
import logging
import copy
# Set up argument processing
parser = argparse.ArgumentParser(description="MHS algorithm correctness tester")
# Suppress log noise from requests library
logging.getLogger("requests").setLevel(logging.WARNING)
# Add arguments
parser.add_argument("algorithm_list_file", help="JSON file of algorithms to benchmark")
parser.add_argument("test_data_file", help="Test input file containing both sets and transversals")
parser.add_argument("errors_file", help="File to write errors in JSON format")
parser.add_argument("-j", dest="num_threads", type=int, nargs='*', help="Number of threads to use for supporting algorithms")
parser.add_argument("-d", dest="docker_base_url", default=None, help="Base URL for Docker client")
parser.add_argument('-v', '--verbose', action="count", default=0, help="Print verbose logs (may be used multiple times)")
parser.add_argument('-s', '--slow', dest="slow", action="store_true", help="Include slow algorithms (be careful!)")
# Process the arguments
args = parser.parse_args()
# Set up logging
log_format = '%(levelname)s [%(asctime)s] %(message)s'
if args.verbose == 0:
log_level = logging.WARNING
elif args.verbose == 1:
log_level = logging.INFO
else:
log_level = logging.DEBUG
logging.basicConfig(format = log_format, level = log_level)
# Read and process files
with open(args.algorithm_list_file) as alg_list_file:
alg_list = json.load(alg_list_file)["containers"]
with open(args.test_data_file) as test_data_file:
test_data_dict = json.load(test_data_file)
correct_transversals = frozenset(frozenset(transversal) for transversal in test_data_dict["transversals"])
# Filter out slow algorithms if requested
if not args.slow:
alg_list = filter(lambda alg: not alg.get("slow"), alg_list)
# Filter out non-threading algorithms if appropriate
num_threads = args.num_threads
if num_threads is None:
num_threads = [1]
if 1 not in num_threads:
alg_list = filter(lambda alg: alg.get("threads"), alg_list)
# Launch containers
logging.info("Launching containers")
alg_collection = pyalgorun.AlgorunContainerCollection(alg_list, docker_base_url = args.docker_base_url)
# Run the tests and store the timing results
tests_failed = []
logging.info("Running algorithms")
for alg in alg_collection:
# Only iterate over cases supported by the algorithm
alg_entry = next(entry for entry in alg_list if entry["algName"] == alg._name)
alg_thread_list = num_threads if alg_entry.get("threads") else [1]
for t in alg_thread_list:
newname = "{0}-t{1}".format(alg._name, t)
try:
result_str = alg.run_alg(args.test_data_file, 600)
result = json.loads(result_str)
except ValueError:
errormessage = "Algorithm {0} returned invalid JSON: {1}".format(newname, result_str)
raise RuntimeError(errormessage)
# Store the transversals as sets in a set for easy equality testing
transversals = frozenset(frozenset(transversal) for transversal in result["transversals"])
# Test the result for correctness and write diagnostics in case of failure
if transversals != correct_transversals:
logging.debug("Algorithm {0} failed!".format(newname))
false_includes = list(list(transversal) for transversal in transversals.difference(correct_transversals))
false_excludes = list(list(transversal) for transversal in correct_transversals.difference(transversals))
errors = {
"false_excludes": false_excludes,
"false_includes": false_includes
}
result = {"algName": newname, "errors": errors}
tests_failed.append(result)
else:
logging.debug("Algorithm {0} passed.".format(newname))
if len(tests_failed) == 0:
logging.warning("All tests succeded.")
else:
result = {"correctTransversals": test_data_dict["transversals"], "algErrors": tests_failed}
logging.warning("{0} tests failed. Dumping results to {1}.".format(len(tests_failed), args.errors_file))
with open(args.errors_file, 'w') as errors_file:
json.dump(result, errors_file, indent=4, separators=(',', ': ')) # Pretty-print the output