mirror of
https://github.com/henrydcase/pqc.git
synced 2024-11-27 01:41:40 +00:00
78a65d6ec9
* Do tests with pytest to run them in parallel * attempt to handle merge commits better for PR test path Similar to how we solved this for travis * Clean up imports * don't run valgrind if not specified slow_test * Fix functest after initializer rename * upload tests results as junit * Upload test-common files since #200 got merged * Catch test results upload failure
275 lines
9.5 KiB
Python
275 lines
9.5 KiB
Python
import atexit
|
|
import functools
|
|
import logging
|
|
import os
|
|
import shutil
|
|
import subprocess
|
|
import sys
|
|
import tempfile
|
|
import unittest
|
|
|
|
import pqclean
|
|
|
|
|
|
@atexit.register
|
|
def cleanup_testcases():
|
|
"""Clean up any remaining isolated test dirs"""
|
|
print("Cleaning up testcases directory",
|
|
file=sys.stderr)
|
|
for dir_ in TEST_TEMPDIRS:
|
|
shutil.rmtree(dir_, ignore_errors=True)
|
|
|
|
|
|
TEST_TEMPDIRS = []
|
|
|
|
|
|
def isolate_test_files(impl_path, test_prefix,
|
|
dir=os.path.join('..', 'testcases')):
|
|
"""Isolates the test files in a separate directory, to help parallelise.
|
|
|
|
Especially Windows is problematic and needs isolation of all test files:
|
|
its build process will create .obj files EVERYWHERE.
|
|
"""
|
|
try:
|
|
os.mkdir(dir)
|
|
except FileExistsError:
|
|
pass
|
|
test_dir = tempfile.mkdtemp(prefix=test_prefix, dir=dir)
|
|
test_dir = os.path.abspath(test_dir)
|
|
TEST_TEMPDIRS.append(test_dir)
|
|
|
|
# Create layers in folder structure
|
|
nested_dir = os.path.join(test_dir, 'crypto_bla')
|
|
os.mkdir(nested_dir)
|
|
nested_dir = os.path.join(nested_dir, 'scheme')
|
|
os.mkdir(nested_dir)
|
|
|
|
# Create test dependencies structure
|
|
os.mkdir(os.path.join(test_dir, 'test'))
|
|
|
|
# the implementation will go here.
|
|
new_impl_dir = os.path.abspath(os.path.join(nested_dir, 'impl'))
|
|
|
|
def initializer():
|
|
"""Isolate the files to be tested"""
|
|
# Copy common files (randombytes.c, aes.c, ...)
|
|
shutil.copytree(
|
|
os.path.join('..', 'common'), os.path.join(test_dir, 'common'))
|
|
# Copy makefiles
|
|
shutil.copy(os.path.join('..', 'test', 'Makefile'),
|
|
os.path.join(test_dir, 'test', 'Makefile'))
|
|
shutil.copy(os.path.join('..', 'test', 'Makefile.Microsoft_nmake'),
|
|
os.path.join(test_dir, 'test', 'Makefile.Microsoft_nmake'))
|
|
# Copy directories with support files
|
|
for d in ['common', 'test_common', 'crypto_sign', 'crypto_kem']:
|
|
shutil.copytree(
|
|
os.path.join('..', 'test', d),
|
|
os.path.join(test_dir, 'test', d)
|
|
)
|
|
|
|
shutil.copytree(impl_path, new_impl_dir)
|
|
|
|
def destructor():
|
|
"""Clean up the isolated files"""
|
|
shutil.rmtree(test_dir)
|
|
|
|
return (test_dir, new_impl_dir, initializer, destructor)
|
|
|
|
|
|
def run_subprocess(command, working_dir='.', env=None, expected_returncode=0):
|
|
"""
|
|
Helper function to run a shell command and report success/failure
|
|
depending on the exit status of the shell command.
|
|
"""
|
|
if env is not None:
|
|
env_ = os.environ.copy()
|
|
env_.update(env)
|
|
env = env_
|
|
|
|
# Note we need to capture stdout/stderr from the subprocess,
|
|
# then print it, which the unittest will then capture and
|
|
# buffer appropriately
|
|
print(working_dir + " > " + " ".join(command))
|
|
result = subprocess.run(
|
|
command,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.STDOUT,
|
|
cwd=working_dir,
|
|
env=env,
|
|
)
|
|
print(result.stdout.decode('utf-8'))
|
|
if expected_returncode is not None:
|
|
assert result.returncode == expected_returncode, \
|
|
"Got unexpected return code {}".format(result.returncode)
|
|
else:
|
|
return (result.returncode, result.stdout.decode('utf-8'))
|
|
return result.stdout.decode('utf-8')
|
|
|
|
|
|
def make(*args, working_dir='.', env=None, expected_returncode=0, **kwargs):
|
|
"""
|
|
Runs a make target in the specified working directory
|
|
|
|
Usage:
|
|
make('clean', 'targetb', SCHEME='bla')
|
|
"""
|
|
if os.name == 'nt':
|
|
make_command = ['nmake', '/f', 'Makefile.Microsoft_nmake',
|
|
'/NOLOGO', '/E']
|
|
# we need SCHEME_UPPERCASE and IMPLEMENTATION_UPPERCASE with nmake
|
|
for envvar in ['IMPLEMENTATION', 'SCHEME']:
|
|
if envvar in kwargs:
|
|
kwargs['{}_UPPERCASE'.format(envvar)] = (
|
|
kwargs[envvar].upper().replace('-', ''))
|
|
else:
|
|
make_command = ['make']
|
|
|
|
return run_subprocess(
|
|
[
|
|
*make_command,
|
|
*['{}={}'.format(k, v) for k, v in kwargs.items()],
|
|
*args,
|
|
],
|
|
working_dir=working_dir,
|
|
env=env,
|
|
expected_returncode=expected_returncode,
|
|
)
|
|
|
|
|
|
def skip_windows(message="This test is not supported on Windows"):
|
|
def wrapper(f):
|
|
@functools.wraps(f)
|
|
def skip_windows(*args, **kwargs):
|
|
raise unittest.SkipTest(message)
|
|
if os.name == 'nt':
|
|
return skip_windows
|
|
else:
|
|
return f
|
|
return wrapper
|
|
|
|
|
|
def slow_test(f):
|
|
@functools.wraps(f)
|
|
def wrapper(*args, **kwargs):
|
|
if ('CI' in os.environ and 'RUN_SLOW' not in os.environ and
|
|
os.environ.get('TRAVIS_EVENT_TYPE') != 'cron'):
|
|
raise unittest.SkipTest("Slow test skipped on CI run")
|
|
return f(*args, **kwargs)
|
|
return wrapper
|
|
|
|
|
|
def ensure_available(executable):
|
|
"""
|
|
Checks if a command is available.
|
|
|
|
If a command MUST be available, because we are in a CI environment,
|
|
raises an AssertionError.
|
|
|
|
In the docker containers, on Travis and on Windows, CI=true is set.
|
|
"""
|
|
path = shutil.which(executable)
|
|
if path:
|
|
return path
|
|
|
|
# Installing clang-tidy on LLVM will be too much of a mess.
|
|
if ((executable == 'clang-tidy' and sys.platform == 'darwin')
|
|
or 'CI' not in os.environ):
|
|
raise unittest.SkipTest(
|
|
"{} is not available on PATH. Install it to run this test.{}"
|
|
.format(executable, "" if not os.name == 'nt'
|
|
else "On Windows, make sure to add it to PATH")
|
|
)
|
|
raise AssertionError("{} not available on CI".format(executable))
|
|
|
|
|
|
def permit_test(testname, *args, **kwargs):
|
|
if len(args) == 0:
|
|
thing = list(kwargs.values())[0]
|
|
else:
|
|
thing = args[0]
|
|
|
|
if 'PQCLEAN_ONLY_TESTS' in os.environ:
|
|
if not(testname.lower() in os.environ['PQCLEAN_ONLY_TESTS'].lower().split(',')):
|
|
return False
|
|
if 'PQCLEAN_SKIP_TESTS' in os.environ:
|
|
if testname.lower() in os.environ['PQCLEAN_SKIP_TESTS'].lower().split(','):
|
|
return False
|
|
|
|
if isinstance(thing, pqclean.Implementation):
|
|
scheme = thing.scheme
|
|
elif isinstance(thing, pqclean.Scheme):
|
|
scheme = thing
|
|
else:
|
|
return True
|
|
|
|
if 'PQCLEAN_ONLY_TYPES' in os.environ:
|
|
if not(scheme.type.lower() in os.environ['PQCLEAN_ONLY_TYPES'].lower().split(',')):
|
|
return False
|
|
if 'PQCLEAN_SKIP_TYPES' in os.environ:
|
|
if scheme.type.lower() in os.environ['PQCLEAN_SKIP_TYPES'].lower().split(','):
|
|
return False
|
|
if 'PQCLEAN_ONLY_SCHEMES' in os.environ:
|
|
if not(scheme.name.lower() in os.environ['PQCLEAN_ONLY_SCHEMES'].lower().split(',')):
|
|
return False
|
|
if 'PQCLEAN_SKIP_SCHEMES' in os.environ:
|
|
if scheme.name.lower() in os.environ['PQCLEAN_SKIP_SCHEMES'].lower().split(','):
|
|
return False
|
|
|
|
if 'PQCLEAN_ONLY_DIFF' in os.environ:
|
|
if shutil.which('git') is not None:
|
|
# if we're on a non-master branch, and the only changes are in schemes,
|
|
# only run tests on those schemes
|
|
branch_result = subprocess.run(
|
|
['git', 'status', '--porcelain=2', '--branch'],
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.STDOUT,
|
|
cwd="..",
|
|
)
|
|
# ensure we're in a working directory
|
|
if branch_result.returncode != 0:
|
|
return True
|
|
# ensure we're not on master branch
|
|
for branch_line in branch_result.stdout.decode('utf-8').splitlines():
|
|
tokens = branch_line.split(' ')
|
|
if tokens[0] == '#' and tokens[1] == 'branch.head':
|
|
if tokens[2] == 'master':
|
|
return True
|
|
# where are there changes?
|
|
diff_result = subprocess.run(
|
|
['git', 'diff', '--name-only', 'origin/master'],
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.STDOUT
|
|
)
|
|
assert diff_result.returncode == 0, \
|
|
"Got unexpected return code {}".format(diff_result.returncode)
|
|
for diff_line in diff_result.stdout.decode('utf-8').splitlines():
|
|
# don't skip test if there are any changes outside schemes
|
|
if (not diff_line.startswith('crypto_kem') and
|
|
not diff_line.startswith('crypto_sign') and
|
|
not diff_line.startswith(os.path.join('test', 'duplicate_consistency'))):
|
|
logging.info("Running all tests as there are changes "
|
|
"outside of schemes")
|
|
return True
|
|
# do test if the scheme in question has been changed
|
|
if diff_line.startswith(thing.path(base='')):
|
|
return True
|
|
# do test if the scheme's duplicate_consistency files have been changed
|
|
if diff_line.startswith(os.path.join('test', 'duplicate_consistency', scheme.name.lower())):
|
|
return True
|
|
# there were no changes outside schemes, and the scheme in question had no diffs
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
def filtered_test(func):
|
|
funcname = func.__name__[len("test_"):]
|
|
|
|
@functools.wraps(func)
|
|
def wrapper(*args, **kwargs):
|
|
if permit_test(funcname, *args, **kwargs):
|
|
return func(*args, **kwargs)
|
|
else:
|
|
raise unittest.SkipTest("Test disabled by filter")
|
|
return wrapper
|