mirror of
https://github.com/RPCS3/llvm.git
synced 2024-12-11 21:56:15 +00:00
d79c539c3b
The script allows the auto-generation of checks for cost model tests to speed up their creation and help improve coverage, which will help a lot with PR36550. If the need arises we can add support for other analyze passes as well, but the cost models was the one I needed to get done - at the moment it just warns that any other analysis mode is unsupported. I've regenerated a couple of x86 test files to show the effect. Differential Revision: https://reviews.llvm.org/D45272 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@329390 91177308-0d34-0410-b5e6-96231b3b80d8
192 lines
6.9 KiB
Python
192 lines
6.9 KiB
Python
#!/usr/bin/env python2.7
|
|
|
|
"""A script to generate FileCheck statements for 'opt' analysis tests.
|
|
|
|
This script is a utility to update LLVM opt analysis test cases with new
|
|
FileCheck patterns. It can either update all of the tests in the file or
|
|
a single test function.
|
|
|
|
Example usage:
|
|
$ update_analyze_test_checks.py --opt=../bin/opt test/foo.ll
|
|
|
|
Workflow:
|
|
1. Make a compiler patch that requires updating some number of FileCheck lines
|
|
in regression test files.
|
|
2. Save the patch and revert it from your local work area.
|
|
3. Update the RUN-lines in the affected regression tests to look canonical.
|
|
Example: "; RUN: opt < %s -analyze -cost-model -S | FileCheck %s"
|
|
4. Refresh the FileCheck lines for either the entire file or select functions by
|
|
running this script.
|
|
5. Commit the fresh baseline of checks.
|
|
6. Apply your patch from step 1 and rebuild your local binaries.
|
|
7. Re-run this script on affected regression tests.
|
|
8. Check the diffs to ensure the script has done something reasonable.
|
|
9. Submit a patch including the regression test diffs for review.
|
|
|
|
A common pattern is to have the script insert complete checking of every
|
|
instruction. Then, edit it down to only check the relevant instructions.
|
|
The script is designed to make adding checks to a test case fast, it is *not*
|
|
designed to be authoratitive about what constitutes a good test!
|
|
"""
|
|
|
|
import argparse
|
|
import itertools
|
|
import os # Used to advertise this file's name ("autogenerated_note").
|
|
import string
|
|
import subprocess
|
|
import sys
|
|
import tempfile
|
|
import re
|
|
|
|
from UpdateTestChecks import common
|
|
|
|
ADVERT = '; NOTE: Assertions have been autogenerated by '
|
|
|
|
# RegEx: this is where the magic happens.
|
|
|
|
IR_FUNCTION_RE = re.compile('^\s*define\s+(?:internal\s+)?[^@]*@([\w-]+)\s*\(')
|
|
|
|
|
|
|
|
|
|
|
|
def main():
|
|
from argparse import RawTextHelpFormatter
|
|
parser = argparse.ArgumentParser(description=__doc__, formatter_class=RawTextHelpFormatter)
|
|
parser.add_argument('-v', '--verbose', action='store_true',
|
|
help='Show verbose output')
|
|
parser.add_argument('--opt-binary', default='opt',
|
|
help='The opt binary used to generate the test case')
|
|
parser.add_argument(
|
|
'--function', help='The function in the test file to update')
|
|
parser.add_argument('tests', nargs='+')
|
|
args = parser.parse_args()
|
|
|
|
autogenerated_note = (ADVERT + 'utils/' + os.path.basename(__file__))
|
|
|
|
opt_basename = os.path.basename(args.opt_binary)
|
|
if (opt_basename != "opt"):
|
|
print >>sys.stderr, 'ERROR: Unexpected opt name: ' + opt_basename
|
|
sys.exit(1)
|
|
|
|
for test in args.tests:
|
|
if args.verbose:
|
|
print >>sys.stderr, 'Scanning for RUN lines in test file: %s' % (test,)
|
|
with open(test) as f:
|
|
input_lines = [l.rstrip() for l in f]
|
|
|
|
raw_lines = [m.group(1)
|
|
for m in [common.RUN_LINE_RE.match(l) for l in input_lines] if m]
|
|
run_lines = [raw_lines[0]] if len(raw_lines) > 0 else []
|
|
for l in raw_lines[1:]:
|
|
if run_lines[-1].endswith("\\"):
|
|
run_lines[-1] = run_lines[-1].rstrip("\\") + " " + l
|
|
else:
|
|
run_lines.append(l)
|
|
|
|
if args.verbose:
|
|
print >>sys.stderr, 'Found %d RUN lines:' % (len(run_lines),)
|
|
for l in run_lines:
|
|
print >>sys.stderr, ' RUN: ' + l
|
|
|
|
prefix_list = []
|
|
for l in run_lines:
|
|
(tool_cmd, filecheck_cmd) = tuple([cmd.strip() for cmd in l.split('|', 1)])
|
|
|
|
if not tool_cmd.startswith(opt_basename + ' '):
|
|
print >>sys.stderr, 'WARNING: Skipping non-%s RUN line: %s' % (opt_basename, l)
|
|
continue
|
|
|
|
if not filecheck_cmd.startswith('FileCheck '):
|
|
print >>sys.stderr, 'WARNING: Skipping non-FileChecked RUN line: ' + l
|
|
continue
|
|
|
|
tool_cmd_args = tool_cmd[len(opt_basename):].strip()
|
|
tool_cmd_args = tool_cmd_args.replace('< %s', '').replace('%s', '').strip()
|
|
|
|
check_prefixes = [item for m in common.CHECK_PREFIX_RE.finditer(filecheck_cmd)
|
|
for item in m.group(1).split(',')]
|
|
if not check_prefixes:
|
|
check_prefixes = ['CHECK']
|
|
|
|
# FIXME: We should use multiple check prefixes to common check lines. For
|
|
# now, we just ignore all but the last.
|
|
prefix_list.append((check_prefixes, tool_cmd_args))
|
|
|
|
func_dict = {}
|
|
for prefixes, _ in prefix_list:
|
|
for prefix in prefixes:
|
|
func_dict.update({prefix: dict()})
|
|
for prefixes, opt_args in prefix_list:
|
|
if args.verbose:
|
|
print >>sys.stderr, 'Extracted opt cmd: ' + opt_basename + ' ' + opt_args
|
|
print >>sys.stderr, 'Extracted FileCheck prefixes: ' + str(prefixes)
|
|
|
|
raw_tool_outputs = common.invoke_tool(args.opt_binary, opt_args, test)
|
|
|
|
# Split analysis outputs by "Printing analysis " declarations.
|
|
for raw_tool_output in re.split(r'Printing analysis ', raw_tool_outputs):
|
|
common.build_function_body_dictionary(
|
|
common.ANALYZE_FUNCTION_RE, common.scrub_body, [],
|
|
raw_tool_output, prefixes, func_dict, args.verbose)
|
|
|
|
is_in_function = False
|
|
is_in_function_start = False
|
|
prefix_set = set([prefix for prefixes, _ in prefix_list for prefix in prefixes])
|
|
if args.verbose:
|
|
print >>sys.stderr, 'Rewriting FileCheck prefixes: %s' % (prefix_set,)
|
|
output_lines = []
|
|
output_lines.append(autogenerated_note)
|
|
|
|
for input_line in input_lines:
|
|
if is_in_function_start:
|
|
if input_line == '':
|
|
continue
|
|
if input_line.lstrip().startswith(';'):
|
|
m = common.CHECK_RE.match(input_line)
|
|
if not m or m.group(1) not in prefix_set:
|
|
output_lines.append(input_line)
|
|
continue
|
|
|
|
# Print out the various check lines here.
|
|
common.add_analyze_checks(output_lines, ';', prefix_list, func_dict, func_name)
|
|
is_in_function_start = False
|
|
|
|
if is_in_function:
|
|
if common.should_add_line_to_output(input_line, prefix_set):
|
|
# This input line of the function body will go as-is into the output.
|
|
# Except make leading whitespace uniform: 2 spaces.
|
|
input_line = common.SCRUB_LEADING_WHITESPACE_RE.sub(r' ', input_line)
|
|
output_lines.append(input_line)
|
|
else:
|
|
continue
|
|
if input_line.strip() == '}':
|
|
is_in_function = False
|
|
continue
|
|
|
|
# Discard any previous script advertising.
|
|
if input_line.startswith(ADVERT):
|
|
continue
|
|
|
|
# If it's outside a function, it just gets copied to the output.
|
|
output_lines.append(input_line)
|
|
|
|
m = IR_FUNCTION_RE.match(input_line)
|
|
if not m:
|
|
continue
|
|
func_name = m.group(1)
|
|
if args.function is not None and func_name != args.function:
|
|
# When filtering on a specific function, skip all others.
|
|
continue
|
|
is_in_function = is_in_function_start = True
|
|
|
|
if args.verbose:
|
|
print>>sys.stderr, 'Writing %d lines to %s...' % (len(output_lines), test)
|
|
|
|
with open(test, 'wb') as f:
|
|
f.writelines([l + '\n' for l in output_lines])
|
|
|
|
|
|
if __name__ == '__main__':
|
|
main()
|