#!/usr/bin/env python3
# A tool to parse ASTMatchers.h and update the documentation in
# ../LibASTMatchersReference.html automatically. Run from the
# directory in which this file is located to update the docs.
import collections
import re
try:
from urllib.request import urlopen
except ImportError:
from urllib2 import urlopen
CLASS_INDEX_PAGE_URL = "https://clang.llvm.org/doxygen/classes.html"
try:
CLASS_INDEX_PAGE = urlopen(CLASS_INDEX_PAGE_URL).read().decode("utf-8")
except Exception as e:
CLASS_INDEX_PAGE = None
print("Unable to get %s: %s" % (CLASS_INDEX_PAGE_URL, e))
MATCHERS_FILE = "../../include/clang/ASTMatchers/ASTMatchers.h"
# Each matcher is documented in one row of the form:
# result | name | argA
# The subsequent row contains the documentation and is hidden by default,
# becoming visible via javascript when the user clicks the matcher name.
TD_TEMPLATE = """
<tr><td>%(result)s</td><td class="name" onclick="toggle('%(id)s')"><a name="%(id)sAnchor">%(name)s</a></td><td>%(args)s</td></tr>
<tr><td colspan="4" class="doc" id="%(id)s"><pre>%(comment)s</pre></td></tr>
"""
# We categorize the matchers into these three categories in the reference:
node_matchers = {}
narrowing_matchers = {}
traversal_matchers = {}
# We output multiple rows per matcher if the matcher can be used on multiple
# node types. Thus, we need a new id per row to control the documentation
# pop-up. ids[name] keeps track of those ids.
ids = collections.defaultdict(int)
# Cache for doxygen urls we have already verified.
doxygen_probes = {}
def esc(text):
"""Escape any html in the given text."""
text = re.sub(r"&", "&", text)
text = re.sub(r"<", "<", text)
text = re.sub(r">", ">", text)
def link_if_exists(m):
"""Wrap a likely AST node name in a link to its clang docs.
We want to do this only if the page exists, in which case it will be
referenced from the class index page.
"""
name = m.group(1)
url = "https://clang.llvm.org/doxygen/classclang_1_1%s.html" % name
if url not in doxygen_probes:
search_str = 'href="classclang_1_1%s.html"' % name
if CLASS_INDEX_PAGE is not None:
doxygen_probes[url] = search_str in CLASS_INDEX_PAGE
else:
doxygen_probes[url] = True
if not doxygen_probes[url]:
print("Did not find %s in class index page" % name)
if doxygen_probes[url]:
return r'Matcher<<a href="%s">%s</a>>' % (url, name)
else:
return m.group(0)
text = re.sub(r"Matcher<([^\*&]+)>", link_if_exists, text)
return text
def extract_result_types(comment):
"""Extracts a list of result types from the given comment.
We allow annotations in the comment of the matcher to specify what
nodes a matcher can match on. Those comments have the form:
Usable as: Any Matcher | (Matcher<T1>[, Matcher<t2>[, ...]])
Returns ['*'] in case of 'Any Matcher', or ['T1', 'T2', ...].
Returns the empty list if no 'Usable as' specification could be
parsed.
"""
result_types = []
m = re.search(r"Usable as: Any Matcher[\s\n]*$", comment, re.S)
if m:
return ["*"]
while True:
m = re.match(r"^(.*)Matcher<([^>]+)>\s*,?[\s\n]*$", comment, re.S)
if not m:
if re.search(r"Usable as:\s*$", comment):
return result_types
else:
return None
result_types += [m.group(2)]
comment = m.group(1)
def strip_doxygen(comment):
"""Returns the given comment without \-escaped words."""
# If there is only a doxygen keyword in the line, delete the whole line.
comment = re.sub(r"^\\[^\s]+\n", r"", comment, flags=re.M)
# If there is a doxygen \see command, change the \see prefix into "See also:".
# FIXME: it would be better to turn this into a link to the target instead.
comment = re.sub(r"\\see", r"See also:", comment)
# Delete the doxygen command and the following whitespace.
comment = re.sub(r"\\[^\s]+\s+", r"", comment)
return comment
def unify_arguments(args):
"""Gets rid of anything the user doesn't care about in the argument list."""
args = re.sub(r"clang::ast_matchers::internal::", r"", args)
args = re.sub(r"ast_matchers::internal::", r"", args)
args = re.sub(r"internal::", r"", args)
args = re.sub(r"extern const\s+(.*)&", r"\1 ", args)
args = re.sub(r"&", r" ", args)
args = re.sub(r"(^|\s)M\d?(\s)", r"\1Matcher<*>\2", args)
args = re.sub(r"BindableMatcher", r"Matcher", args)
args = re.sub(r"const Matcher", r"Matcher", args)
return args
def unify_type(result_type):
"""Gets rid of anything the user doesn't care about in the type name."""
result_type = re.sub(
r"^internal::(Bindable)?Matcher<([a-zA-Z_][a-zA-Z0-9_]*)>$", r"\2", result_type
)
return result_type
def add_matcher(result_type, name, args, comment, is_dyncast=False):
"""Adds a matcher to one of our categories."""
if name == "id":
# FIXME: Figure out whether we want to support the 'id' matcher.
return
matcher_id = "%s%d" % (name, ids[name])
ids[name] += 1
args = unify_arguments(args)
result_type = unify_type(result_type)
docs_result_type = esc("Matcher<%s>" % result_type)
if name == "mapAnyOf":
args = "nodeMatcherFunction..."
docs_result_type = "<em>unspecified</em>"
matcher_html = TD_TEMPLATE % {
"result": docs_result_type,
"name": name,
"args": esc(args),
"comment": esc(strip_doxygen(comment)),
"id": matcher_id,
}
if is_dyncast:
dict = node_matchers
lookup = result_type + name
# Use a heuristic to figure out whether a matcher is a narrowing or
# traversal matcher. By default, matchers that take other matchers as
# arguments (and are not node matchers) do traversal. We specifically
# exclude known narrowing matchers that also take other matchers as
# arguments.
elif "Matcher<" not in args or name in [
"allOf",
"anyOf",
"anything",
"unless",
"mapAnyOf",
]:
dict = narrowing_matchers
lookup = result_type + name + esc(args)
else:
dict = traversal_matchers
lookup = result_type + name + esc(args)
if dict.get(lookup) is None or len(dict.get(lookup)) < len(matcher_html):
dict[lookup] = matcher_html
def act_on_decl(declaration, comment, allowed_types):
"""Parse the matcher out of the given declaration and comment.
If 'allowed_types' is set, it contains a list of node types the matcher
can match on, as extracted from the static type asserts in the matcher
definition.
"""
if declaration.strip():
if re.match(r"^\s?(#|namespace|using|template <typename NodeType> using|})", declaration):
return
# Node matchers are defined by writing:
# VariadicDynCastAllOfMatcher<ResultType, ArgumentType> name;
m = re.match(
r""".*Variadic(?:DynCast)?AllOfMatcher\s*<
\s*([^\s,]+)\s*(?:,
\s*([^\s>]+)\s*)?>
\s*([^\s;]+)\s*;\s*$""",
declaration,
flags=re.X,
)
if m:
result, inner, name = m.groups()
if not inner:
inner = result
add_matcher(
result, name, "Matcher<%s>..." % inner, comment, is_dyncast=True
)
return
# Special case of type matchers:
# AstTypeMatcher<ArgumentType> name
m = re.match(
r""".*AstTypeMatcher\s*<
\s*([^\s>]+)\s*>
\s*([^\s;]+)\s*;\s*$""",
declaration,
flags=re.X,
)
if m:
inner, name = m.groups()
add_matcher(
"Type", name, "Matcher<%s>..." % inner, comment, is_dyncast=True
)
# FIXME: re-enable once we have implemented casting on the TypeLoc
# hierarchy.
# add_matcher('TypeLoc', '%sLoc' % name, 'Matcher<%sLoc>...' % inner,
# comment, is_dyncast=True)
return
# Parse the various matcher definition macros.
m = re.match(
""".*AST_TYPE(LOC)?_TRAVERSE_MATCHER(?:_DECL)?\(
\s*([^\s,]+\s*),
\s*(?:[^\s,]+\s*),
\s*AST_POLYMORPHIC_SUPPORTED_TYPES\(([^)]*)\)
\)\s*;\s*$""",
declaration,
flags=re.X,
)
if m:
loc, name, results = m.groups()[0:3]
result_types = [r.strip() for r in results.split(",")]
comment_result_types = extract_result_types(comment)
if comment_result_types and sorted(result_types) != sorted(
comment_result_types
):
raise Exception("Inconsistent documentation for: %s" % name)
for result_type in result_types:
add_matcher(result_type, name, "Matcher<Type>", comment)
# if loc:
# add_matcher('%sLoc' % result_type, '%sLoc' % name, 'Matcher<TypeLoc>',
# comment)
return
m = re.match(
r"""^\s*AST_POLYMORPHIC_MATCHER(_P)?(.?)(?:_OVERLOAD)?\(
\s*([^\s,]+)\s*,
\s*AST_POLYMORPHIC_SUPPORTED_TYPES\(([^)]*)\)
(?:,\s*([^\s,]+)\s*
,\s*([^\s,]+)\s*)?
(?:,\s*([^\s,]+)\s*
,\s*([^\s,]+)\s*)?
(?:,\s*\d+\s*)?
\)\s*{\s*$""",
declaration,
flags=re.X,
)
if m:
p, n, name, results = m.groups()[0:4]
args = m.groups()[4:]
result_types = [r.strip() for r in results.split(",")]
if allowed_types and allowed_types != result_types:
raise Exception("Inconsistent documentation for: %s" % name)
if n not in ["", "2"]:
raise Exception('Cannot parse "%s"' % declaration)
args = ", ".join(
"%s %s" % (args[i], args[i + 1])
for i in range(0, len(args), 2)
if args[i]
)
for result_type in result_types:
add_matcher(result_type, name, args, comment)
return
m = re.match(
r"""^\s*AST_POLYMORPHIC_MATCHER_REGEX(?:_OVERLOAD)?\(
\s*([^\s,]+)\s*,
\s*AST_POLYMORPHIC_SUPPORTED_TYPES\(([^)]*)\),
\s*([^\s,]+)\s*
(?:,\s*\d+\s*)?
\)\s*{\s*$""",
declaration,
flags=re.X,
)
if m:
name, results, arg_name = m.groups()[0:3]
result_types = [r.strip() for r in results.split(",")]
if allowed_types and allowed_types != result_types:
raise Exception("Inconsistent documentation for: %s" % name)
arg = "StringRef %s, Regex::RegexFlags Flags = NoFlags" % arg_name
comment += """
If the matcher is used in clang-query, RegexFlags parameter
should be passed as a quoted string. e.g: "NoFlags".
Flags can be combined with '|' example \"IgnoreCase | BasicRegex\"
"""
for result_type in result_types:
add_matcher(result_type, name, arg, comment)
return
m = re.match(
r"""^\s*AST_MATCHER_FUNCTION(_P)?(.?)(?:_OVERLOAD)?\(
(?:\s*([^\s,]+)\s*,)?
\s*([^\s,]+)\s*
(?:,\s*([^\s,]+)\s*
,\s*([^\s,]+)\s*)?
(?:,\s*([^\s,]+)\s*
,\s*([^\s,]+)\s*)?
(?:,\s*\d+\s*)?
\)\s*{\s*$""",
declaration,
flags=re.X,
)
if m:
p, n, result, name = m.groups()[0:4]
args = m.groups()[4:]
if n not in ["", "2"]:
raise Exception('Cannot parse "%s"' % declaration)
args = ", ".join(
"%s %s" % (args[i], args[i + 1])
for i in range(0, len(args), 2)
if args[i]
)
add_matcher(result, name, args, comment)
return
m = re.match(
r"""^\s*AST_MATCHER(_P)?(.?)(?:_OVERLOAD)?\(
(?:\s*([^\s,]+)\s*,)?
\s*([^\s,]+)\s*
(?:,\s*([^,]+)\s*
,\s*([^\s,]+)\s*)?
(?:,\s*([^\s,]+)\s*
,\s*([^\s,]+)\s*)?
(?:,\s*\d+\s*)?
\)\s*{""",
declaration,
flags=re.X,
)
if m:
p, n, result, name = m.groups()[0:4]
args = m.groups()[4:]
if not result:
if not allowed_types:
raise Exception("Did not find allowed result types for: %s" % name)
result_types = allowed_types
else:
result_types = [result]
if n not in ["", "2"]:
raise Exception('Cannot parse "%s"' % declaration)
args = ", ".join(
"%s %s" % (args[i], args[i + 1])
for i in range(0, len(args), 2)
if args[i]
)
for result_type in result_types:
add_matcher(result_type, name, args, comment)
return
m = re.match(
r"""^\s*AST_MATCHER_REGEX(?:_OVERLOAD)?\(
\s*([^\s,]+)\s*,
\s*([^\s,]+)\s*,
\s*([^\s,]+)\s*
(?:,\s*\d+\s*)?
\)\s*{""",
declaration,
flags=re.X,
)
if m:
result, name, arg_name = m.groups()[0:3]
if not result:
if not allowed_types:
raise Exception("Did not find allowed result types for: %s" % name)
result_types = allowed_types
else:
result_types = [result]
arg = "StringRef %s, Regex::RegexFlags Flags = NoFlags" % arg_name
comment += """
If the matcher is used in clang-query, RegexFlags parameter
should be passed as a quoted string. e.g: "NoFlags".
Flags can be combined with '|' example \"IgnoreCase | BasicRegex\"
"""
for result_type in result_types:
add_matcher(result_type, name, arg, comment)
return
# Parse ArgumentAdapting matchers.
m = re.match(
r"""^.*ArgumentAdaptingMatcherFunc<.*>\s*
([a-zA-Z]*);$""",
declaration,
flags=re.X,
)
if m:
name = m.groups()[0]
add_matcher("*", name, "Matcher<*>", comment)
return
# Parse Variadic functions.
m = re.match(
r"""^.*internal::VariadicFunction\s*<\s*([^,]+),\s*([^,]+),\s*[^>]+>\s*
([a-zA-Z]*);$""",
declaration,
flags=re.X,
)
if m:
result, arg, name = m.groups()[:3]
add_matcher(result, name, "%s, ..., %s" % (arg, arg), comment)
return
m = re.match(
r"""^.*internal::VariadicFunction\s*<\s*
internal::PolymorphicMatcher<[\S\s]+
AST_POLYMORPHIC_SUPPORTED_TYPES\(([^)]*)\),\s*(.*);$""",
declaration,
flags=re.X,
)
if m:
results, trailing = m.groups()
trailing, name = trailing.rsplit(">", 1)
name = name.strip()
trailing, _ = trailing.rsplit(",", 1)
_, arg = trailing.rsplit(",", 1)
arg = arg.strip()
result_types = [r.strip() for r in results.split(",")]
for result_type in result_types:
add_matcher(result_type, name, "%s, ..., %s" % (arg, arg), comment)
return
# Parse Variadic operator matchers.
m = re.match(
r"""^.*VariadicOperatorMatcherFunc\s*<\s*([^,]+),\s*([^\s]+)\s*>\s*
([a-zA-Z]*);$""",
declaration,
flags=re.X,
)
if m:
min_args, max_args, name = m.groups()[:3]
if max_args == "1":
add_matcher("*", name, "Matcher<*>", comment)
return
elif max_args == "std::numeric_limits<unsigned>::max()":
add_matcher("*", name, "Matcher<*>, ..., Matcher<*>", comment)
return
m = re.match(
r"""^.*MapAnyOfMatcher<.*>\s*
([a-zA-Z]*);$""",
declaration,
flags=re.X,
)
if m:
name = m.groups()[0]
add_matcher("*", name, "Matcher<*>...Matcher<*>", comment)
return
# Parse free standing matcher functions, like:
# Matcher<ResultType> Name(Matcher<ArgumentType> InnerMatcher) {
m = re.match(
r"""^\s*(?:template\s+<\s*(?:class|typename)\s+(.+)\s*>\s+)?
(.*)\s+
([^\s\(]+)\s*\(
(.*)
\)\s*{""",
declaration,
re.X,
)
if m:
template_name, result, name, args = m.groups()
if template_name:
matcherTemplateArgs = re.findall(
r"Matcher<\s*(%s)\s*>" % template_name, args
)
templateArgs = re.findall(
r"(?:^|[\s,<])(%s)(?:$|[\s,>])" % template_name, args
)
if len(matcherTemplateArgs) < len(templateArgs):
# The template name is used naked, so don't replace with `*`` later on
template_name = None
else:
args = re.sub(
r"(^|[\s,<])%s($|[\s,>])" % template_name, r"\1*\2", args
)
args = ", ".join(p.strip() for p in args.split(","))
m = re.match(r"(?:^|.*\s+)internal::(?:Bindable)?Matcher<([^>]+)>$", result)
if m:
result_types = [m.group(1)]
if (
template_name
and len(result_types) == 1
and result_types[0] == template_name
):
result_types = ["*"]
else:
result_types = extract_result_types(comment)
if not result_types:
if not comment:
# Only overloads don't have their own doxygen comments; ignore those.
print('Ignoring "%s"' % name)
else:
print('Cannot determine result type for "%s"' % name)
else:
for result_type in result_types:
add_matcher(result_type, name, args, comment)
else:
print('*** Unparsable: "' + declaration + '" ***')
def sort_table(matcher_type, matcher_map):
"""Returns the sorted html table for the given row map."""
table = ""
for key in sorted(matcher_map.keys()):
table += matcher_map[key] + "\n"
return (
"<!-- START_%(type)s_MATCHERS -->\n"
+ "%(table)s"
+ "<!--END_%(type)s_MATCHERS -->"
) % {
"type": matcher_type,
"table": table,
}
# Parse the ast matchers.
# We alternate between two modes:
# body = True: We parse the definition of a matcher. We need
# to parse the full definition before adding a matcher, as the
# definition might contain static asserts that specify the result
# type.
# body = False: We parse the comments and declaration of the matcher.
comment = ""
declaration = ""
allowed_types = []
body = False
for line in open(MATCHERS_FILE).read().splitlines():
if body:
if line.strip() and line[0] == "}":
if declaration:
act_on_decl(declaration, comment, allowed_types)
comment = ""
declaration = ""
allowed_types = []
body = False
else:
m = re.search(r"is_base_of<([^,]+), NodeType>", line)
if m and m.group(1):
allowed_types += [m.group(1)]
continue
if line.strip() and line.lstrip()[0] == "/":
comment += re.sub(r"^/+\s?", "", line) + "\n"
else:
declaration += " " + line
if (
(not line.strip())
or line.rstrip()[-1] == ";"
or (line.rstrip()[-1] == "{" and line.rstrip()[-3:] != "= {")
):
if line.strip() and line.rstrip()[-1] == "{":
body = True
else:
act_on_decl(declaration, comment, allowed_types)
comment = ""
declaration = ""
allowed_types = []
node_matcher_table = sort_table("DECL", node_matchers)
narrowing_matcher_table = sort_table("NARROWING", narrowing_matchers)
traversal_matcher_table = sort_table("TRAVERSAL", traversal_matchers)
reference = open("../LibASTMatchersReference.html").read()
reference = re.sub(
r"<!-- START_DECL_MATCHERS.*END_DECL_MATCHERS -->",
node_matcher_table,
reference,
flags=re.S,
)
reference = re.sub(
r"<!-- START_NARROWING_MATCHERS.*END_NARROWING_MATCHERS -->",
narrowing_matcher_table,
reference,
flags=re.S,
)
reference = re.sub(
r"<!-- START_TRAVERSAL_MATCHERS.*END_TRAVERSAL_MATCHERS -->",
traversal_matcher_table,
reference,
flags=re.S,
)
with open("../LibASTMatchersReference.html", "w", newline="\n") as output:
output.write(reference)