From: Florian Weimer Date: Mon, 17 Feb 2020 16:18:49 +0000 (+0100) Subject: conform/conformtest.py: Extend tokenizer to cover character constants X-Git-Tag: upstream/2.34~2210 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=1f1c65e232878bca88fe8551318c7bc02b90ee90;p=platform%2Fupstream%2Fglibc.git conform/conformtest.py: Extend tokenizer to cover character constants Such constants are used in __USE_EXTERN_INLINES blocks. --- diff --git a/conform/conformtest.py b/conform/conformtest.py index 951e3b2..cb2bd97 100644 --- a/conform/conformtest.py +++ b/conform/conformtest.py @@ -633,12 +633,11 @@ class HeaderTests(object): bad_tokens.discard(match.group(1)) continue # Tokenize the line and check identifiers found. The - # handling of strings does not allow for escaped - # quotes, no allowance is made for character - # constants, and hex floats may be wrongly split into - # tokens including identifiers, but this is sufficient - # in practice and matches the old perl script. - line = re.sub(r'"[^"]*"', '', line) + # handling of strings and character constants does not + # allow for escaped quotes, and hex floats may be + # wrongly split into tokens including identifiers, but + # this is sufficient in practice. + line = re.sub(r'(?:\bL)?(?:"[^"]*"|\'[^\']*\')', '', line) line = line.strip() for token in re.split(r'[^A-Za-z0-9_]+', line): if re.match(r'[A-Za-z_]', token):