2
0
mirror of https://github.com/openvswitch/ovs synced 2025-08-22 09:58:01 +00:00
ovs/utilities/checkpatch.py
Chandan Somani d25c6bd8df checkpatch: Reorganize flagged words using a list.
Single out flagged words and allow for more useful
details, like spelling suggestions.

Signed-off-by: Chandan Somani <csomani@redhat.com>
Signed-off-by: Eelco Chaudron <echaudro@redhat.com>
2023-07-12 12:01:09 +02:00

1134 lines
41 KiB
Python
Executable File

#!/usr/bin/env python3
# Copyright (c) 2016, 2017 Red Hat, Inc.
# Copyright (c) 2018 Nicira, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at:
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import email
import getopt
import os
import re
import sys
RETURN_CHECK_INITIAL_STATE = 0
RETURN_CHECK_STATE_WITH_RETURN = 1
RETURN_CHECK_AWAITING_BRACE = 2
EXIT_FAILURE = 1
__errors = 0
__warnings = 0
empty_return_check_state = 0
print_file_name = None
checking_file = False
total_line = 0
colors = False
spellcheck = False
quiet = False
spell_check_dict = None
def open_spell_check_dict():
import enchant
try:
extra_keywords = ['ovs', 'vswitch', 'vswitchd', 'ovs-vswitchd',
'netdev', 'selinux', 'ovs-ctl', 'dpctl', 'ofctl',
'openvswitch', 'dpdk', 'hugepage', 'hugepages',
'pmd', 'upcall', 'vhost', 'rx', 'tx', 'vhostuser',
'openflow', 'qsort', 'rxq', 'txq', 'perf', 'stats',
'struct', 'int', 'char', 'bool', 'upcalls', 'nicira',
'bitmask', 'ipv4', 'ipv6', 'tcp', 'tcp4', 'tcpv4',
'udp', 'udp4', 'udpv4', 'icmp', 'icmp4', 'icmpv6',
'vlan', 'vxlan', 'cksum', 'csum', 'checksum',
'ofproto', 'numa', 'mempool', 'mempools', 'mbuf',
'mbufs', 'hmap', 'cmap', 'smap', 'dhcpv4', 'dhcp',
'dhcpv6', 'opts', 'metadata', 'geneve', 'mutex',
'netdev', 'netdevs', 'subtable', 'virtio', 'qos',
'policer', 'datapath', 'tunctl', 'attr', 'ethernet',
'ether', 'defrag', 'defragment', 'loopback', 'sflow',
'acl', 'initializer', 'recirc', 'xlated', 'unclosed',
'netlink', 'msec', 'usec', 'nsec', 'ms', 'us', 'ns',
'kilobits', 'kbps', 'kilobytes', 'megabytes', 'mbps',
'gigabytes', 'gbps', 'megabits', 'gigabits', 'pkts',
'tuple', 'miniflow', 'megaflow', 'conntrack',
'vlans', 'vxlans', 'arg', 'tpid', 'xbundle',
'xbundles', 'mbundle', 'mbundles', 'netflow',
'localnet', 'odp', 'pre', 'dst', 'dest', 'src',
'ethertype', 'cvlan', 'ips', 'msg', 'msgs',
'liveness', 'userspace', 'eventmask', 'datapaths',
'slowpath', 'fastpath', 'multicast', 'unicast',
'revalidation', 'namespace', 'qdisc', 'uuid',
'ofport', 'subnet', 'revalidation', 'revalidator',
'revalidate', 'l2', 'l3', 'l4', 'openssl', 'mtu',
'ifindex', 'enum', 'enums', 'http', 'https', 'num',
'vconn', 'vconns', 'conn', 'nat', 'memset', 'memcmp',
'strcmp', 'strcasecmp', 'tc', 'ufid', 'api',
'ofpbuf', 'ofpbufs', 'hashmaps', 'hashmap', 'deref',
'dereference', 'hw', 'prio', 'sendmmsg', 'sendmsg',
'malloc', 'free', 'alloc', 'pid', 'ppid', 'pgid',
'uid', 'gid', 'sid', 'utime', 'stime', 'cutime',
'cstime', 'vsize', 'rss', 'rsslim', 'whcan', 'gtime',
'eip', 'rip', 'cgtime', 'dbg', 'gw', 'sbrec', 'bfd',
'sizeof', 'pmds', 'nic', 'nics', 'hwol', 'encap',
'decap', 'tlv', 'tlvs', 'decapsulation', 'fd',
'cacheline', 'xlate', 'skiplist', 'idl',
'comparator', 'natting', 'alg', 'pasv', 'epasv',
'wildcard', 'nated', 'amd64', 'x86_64',
'recirculation', 'linux', 'afxdp', 'promisc', 'goto',
'misconfigured', 'misconfiguration', 'checkpatch',
'debian', 'travis', 'cirrus', 'appveyor', 'faq',
'erspan', 'const', 'hotplug', 'addresssanitizer',
'ovsdb', 'dpif', 'veth', 'rhel', 'jsonrpc', 'json',
'syscall', 'lacp', 'ipf', 'skb', 'valgrind']
global spell_check_dict
spell_check_dict = enchant.Dict("en_US")
for kw in extra_keywords:
spell_check_dict.add(kw)
return True
except:
return False
def get_color_end():
global colors
if colors:
return "\033[00m"
return ""
def get_red_begin():
global colors
if colors:
return "\033[91m"
return ""
def get_yellow_begin():
global colors
if colors:
return "\033[93m"
return ""
def print_error(message):
global __errors
print("%sERROR%s: %s" % (get_red_begin(), get_color_end(), message))
__errors = __errors + 1
def print_warning(message):
global __warnings
print("%sWARNING%s: %s" % (get_yellow_begin(), get_color_end(), message))
__warnings = __warnings + 1
def reset_counters():
global __errors, __warnings, total_line
__errors = 0
__warnings = 0
total_line = 0
# These are keywords whose names are normally followed by a space and
# something in parentheses (usually an expression) then a left curly brace.
#
# 'do' almost qualifies but it's also used as "do { ... } while (...);".
__parenthesized_constructs = 'if|for|while|switch|[_A-Z]+FOR_*EACH[_A-Z0-9]*'
__regex_added_line = re.compile(r'^\+{1,2}[^\+][\w\W]*')
__regex_subtracted_line = re.compile(r'^\-{1,2}[^\-][\w\W]*')
__regex_leading_with_whitespace_at_all = re.compile(r'^\s+')
__regex_leading_with_spaces = re.compile(r'^ +[\S]+')
__regex_trailing_whitespace = re.compile(r'[^\S]+$')
__regex_single_line_feed = re.compile(r'^\f$')
__regex_for_if_missing_whitespace = re.compile(r' +(%s)[\(]'
% __parenthesized_constructs)
__regex_hash_define_for_each = re.compile(
r'#define [_A-Z]+FOR_*EACH[_A-Z0-9]*\(')
__regex_for_if_too_much_whitespace = re.compile(r' +(%s) +[\(]'
% __parenthesized_constructs)
__regex_for_if_parens_whitespace = \
re.compile(r' +(%s) \( +[\s\S]+\)' % __parenthesized_constructs)
__regex_is_for_if_single_line_bracket = \
re.compile(r'^ +(%s) \(.*\)' % __parenthesized_constructs)
__regex_ends_with_bracket = \
re.compile(r'[^\s]\) {(\s+/\*[\s\Sa-zA-Z0-9\.,\?\*/+-]*)?$')
__regex_ptr_declaration_missing_whitespace = re.compile(r'[a-zA-Z0-9]\*[^*]')
__regex_cast_missing_whitespace = re.compile(r'\)[a-zA-Z0-9]')
__regex_is_comment_line = re.compile(r'^\s*(/\*|\*\s)')
__regex_has_comment = re.compile(r'.*(/\*|\*\s)')
__regex_has_c99_comment = re.compile(r'.*//.*$')
__regex_trailing_operator = re.compile(r'^[^ ]* [^ ]*[?:]$')
__regex_conditional_else_bracing = re.compile(r'^\s*else\s*{?$')
__regex_conditional_else_bracing2 = re.compile(r'^\s*}\selse\s*$')
__regex_has_xxx_mark = re.compile(r'.*xxx.*', re.IGNORECASE)
__regex_added_doc_rst = re.compile(
r'\ndiff .*Documentation/.*rst\nnew file mode')
__regex_empty_return = re.compile(r'\s*return;')
__regex_if_macros = re.compile(r'^ +(%s) \([\S]([\s\S]+[\S])*\) { +\\' %
__parenthesized_constructs)
__regex_nonascii_characters = re.compile("[^\u0000-\u007f]")
__regex_efgrep = re.compile(r'.*[ef]grep.*$')
skip_leading_whitespace_check = False
skip_trailing_whitespace_check = False
skip_gerrit_change_id_check = False
skip_block_whitespace_check = False
skip_signoff_check = False
# Don't enforce character limit on files that include these characters in their
# name, as they may have legitimate reasons to have longer lines.
#
# Python isn't checked as flake8 performs these checks during build.
line_length_ignore_list = re.compile(
r'\.(am|at|etc|in|m4|mk|patch|py|yml)$|^debian/.*$')
# Don't enforce a requirement that leading whitespace be all spaces on
# files that include these characters in their name, since these kinds
# of files need lines with leading tabs.
leading_whitespace_ignore_list = re.compile(r'\.(mk|am|at)$|^debian/.*$')
def is_subtracted_line(line):
"""Returns TRUE if the line in question has been removed."""
return __regex_subtracted_line.search(line) is not None
def is_added_line(line):
"""Returns TRUE if the line in question is an added line.
"""
global checking_file
return __regex_added_line.search(line) is not None or checking_file
def added_line(line):
"""Returns the line formatted properly by removing diff syntax"""
global checking_file
if not checking_file:
return line[1:]
return line
def leading_whitespace_is_spaces(line):
"""Returns TRUE if the leading whitespace in added lines is spaces
"""
if skip_leading_whitespace_check:
return True
if (__regex_leading_with_whitespace_at_all.search(line) is not None and
__regex_single_line_feed.search(line) is None):
return __regex_leading_with_spaces.search(line) is not None
return True
def trailing_whitespace_or_crlf(line):
"""Returns TRUE if the trailing characters is whitespace
"""
if skip_trailing_whitespace_check:
return False
return (__regex_trailing_whitespace.search(line) is not None and
__regex_single_line_feed.search(line) is None)
def if_and_for_whitespace_checks(line):
"""Return TRUE if there is appropriate whitespace after if, for, while
"""
if skip_block_whitespace_check:
return True
if (__regex_for_if_missing_whitespace.search(line) is not None and
__regex_hash_define_for_each.search(line) is None):
return False
if (__regex_for_if_too_much_whitespace.search(line) is not None or
__regex_for_if_parens_whitespace.search(line)):
return False
return True
def if_and_for_end_with_bracket_check(line):
"""Return TRUE if there is not a bracket at the end of an if, for, while
block which fits on a single line ie: 'if (foo)'"""
def balanced_parens(line):
"""This is a rather naive counter - it won't deal with quotes"""
balance = 0
for letter in line:
if letter == '(':
balance += 1
elif letter == ')':
balance -= 1
return balance == 0
if __regex_is_for_if_single_line_bracket.search(line) is not None:
if not balanced_parens(line):
return True
if __regex_ends_with_bracket.search(line) is None:
if line.endswith("\\") and \
__regex_if_macros.match(line) is not None:
return True
else:
return False
if __regex_conditional_else_bracing.match(line) is not None:
return False
if __regex_conditional_else_bracing2.match(line) is not None:
return False
return True
def pointer_whitespace_check(line):
"""Return TRUE if there is no space between a pointer name and the
asterisk that denotes this is a apionter type, ie: 'struct foo*'"""
return __regex_ptr_declaration_missing_whitespace.search(line) is not None
def nonascii_character_check(line):
"""Return TRUE if inappropriate Unicode characters are detected """
return __regex_nonascii_characters.search(line) is not None
def cast_whitespace_check(line):
"""Return TRUE if there is no space between the '()' used in a cast and
the expression whose type is cast, i.e.: '(void *)foo'"""
return __regex_cast_missing_whitespace.search(line) is not None
def line_length_check(line):
"""Return TRUE if the line length is too long"""
if len(line) > 79:
print_warning("Line is %d characters long (recommended limit is 79)"
% len(line))
return True
return False
def is_comment_line(line):
"""Returns TRUE if the current line is part of a block comment."""
return __regex_is_comment_line.match(line) is not None
def has_comment(line):
"""Returns TRUE if the current line contains a comment or is part of
a block comment."""
return __regex_has_comment.match(line) is not None
def has_c99_comment(line):
"""Returns TRUE if the current line contains C99 style comment (//)."""
return __regex_has_c99_comment.match(line) is not None
def trailing_operator(line):
"""Returns TRUE if the current line ends with an operatorsuch as ? or :"""
return __regex_trailing_operator.match(line) is not None
def has_xxx_mark(line):
"""Returns TRUE if the current line contains 'xxx'."""
return __regex_has_xxx_mark.match(line) is not None
def has_efgrep(line):
"""Returns TRUE if the current line contains 'egrep' or 'fgrep'."""
return __regex_efgrep.match(line) is not None
def filter_comments(current_line, keep=False):
"""remove all of the c-style comments in a line"""
STATE_NORMAL = 0
STATE_COMMENT_SLASH = 1
STATE_COMMENT_CONTENTS = 3
STATE_COMMENT_END_SLASH = 4
state = STATE_NORMAL
sanitized_line = ''
check_state = STATE_NORMAL
only_whitespace = True
if keep:
check_state = STATE_COMMENT_CONTENTS
for c in current_line:
if c == '/':
if state == STATE_NORMAL:
state = STATE_COMMENT_SLASH
elif state == STATE_COMMENT_SLASH:
# This is for c++ style comments. We will warn later
return sanitized_line[:1]
elif state == STATE_COMMENT_END_SLASH:
c = ''
state = STATE_NORMAL
elif c == '*':
if only_whitespace:
# just assume this is a continuation from the previous line
# as a comment
state = STATE_COMMENT_END_SLASH
elif state == STATE_COMMENT_SLASH:
state = STATE_COMMENT_CONTENTS
sanitized_line = sanitized_line[:-1]
elif state == STATE_COMMENT_CONTENTS:
state = STATE_COMMENT_END_SLASH
elif state == STATE_COMMENT_END_SLASH:
# Need to re-introduce the star from the previous state, since
# it may have been clipped by the state check below.
c = '*' + c
state = STATE_COMMENT_CONTENTS
elif state == STATE_COMMENT_SLASH:
# Need to re-introduce the slash from the previous state, since
# it may have been clipped by the state check below.
c = '/' + c
state = STATE_NORMAL
if state != check_state:
c = ''
if not c.isspace():
only_whitespace = False
sanitized_line += c
return sanitized_line
def check_spelling(line, comment):
if not spell_check_dict or not spellcheck:
return False
words = filter_comments(line, True) if comment else line
words = words.replace(':', ' ').split(' ')
flagged_words = []
for word in words:
skip = False
strword = re.subn(r'\W+', '', word)[0].replace(',', '')
if (len(strword)
and not spell_check_dict.check(strword.lower())
and not spell_check_dict.check(word.lower())):
if any([check_char in word
for check_char in ['=', '(', '-', '_', '/', '\'']]):
skip = True
# special case the '.'
if '.' in word and not word.endswith('.'):
skip = True
# skip proper nouns and references to macros
if strword.isupper() or (strword[0].isupper() and
strword[1:].islower()):
skip = True
# skip words containing numbers
if any(check_char.isdigit() for check_char in strword):
skip = True
if not skip:
flagged_words.append(strword)
if len(flagged_words) > 0:
for mistake in flagged_words:
print_warning("Possible misspelled word: \"%s\"" % mistake)
return True
return False
def __check_doc_is_listed(text, doctype, docdir, docfile):
if doctype == 'rst':
beginre = re.compile(r'\+\+\+.*{}/index.rst'.format(docdir))
docre = re.compile(r'\n\+.*{}'.format(docfile.replace('.rst', '')))
elif doctype == 'automake':
beginre = re.compile(r'\+\+\+.*Documentation/automake.mk')
docre = re.compile(r'\n\+\t{}/{}'.format(docdir, docfile))
else:
raise NotImplementedError("Invalid doctype: {}".format(doctype))
res = beginre.search(text)
if res is None:
return True
hunkstart = res.span()[1]
hunkre = re.compile(r'\n(---|\+\+\+) (\S+)')
res = hunkre.search(text[hunkstart:])
if res is None:
hunkend = len(text)
else:
hunkend = hunkstart + res.span()[0]
hunk = text[hunkstart:hunkend]
# find if the file is being added.
if docre.search(hunk) is not None:
return False
return True
def __check_new_docs(text, doctype):
"""Check if the documentation is listed properly. If doctype is 'rst' then
the index.rst is checked. If the doctype is 'automake' then automake.mk
is checked. Returns TRUE if the new file is not listed."""
failed = False
new_docs = __regex_added_doc_rst.findall(text)
for doc in new_docs:
docpathname = doc.split(' ')[2]
gitdocdir, docfile = os.path.split(docpathname.rstrip('\n'))
if docfile == "index.rst":
continue
if gitdocdir.startswith('a/'):
docdir = gitdocdir.replace('a/', '', 1)
else:
docdir = gitdocdir
if __check_doc_is_listed(text, doctype, docdir, docfile):
if doctype == 'rst':
print_warning("New doc {} not listed in {}/index.rst".format(
docfile, docdir))
elif doctype == 'automake':
print_warning("New doc {} not listed in "
"Documentation/automake.mk".format(docfile))
else:
raise NotImplementedError("Invalid doctype: {}".format(
doctype))
failed = True
return failed
def check_doc_docs_automake(text):
return __check_new_docs(text, 'automake')
def check_new_docs_index(text):
return __check_new_docs(text, 'rst')
def empty_return_with_brace(line):
"""Returns TRUE if a function contains a return; followed
by one or more line feeds and terminates with a '}'
at start of line"""
def empty_return(line):
"""Returns TRUE if a function has a 'return;'"""
return __regex_empty_return.match(line) is not None
global empty_return_check_state
if empty_return_check_state == RETURN_CHECK_INITIAL_STATE \
and empty_return(line):
empty_return_check_state = RETURN_CHECK_STATE_WITH_RETURN
elif empty_return_check_state == RETURN_CHECK_STATE_WITH_RETURN \
and (re.match(r'^}$', line) or len(line) == 0):
if re.match('^}$', line):
empty_return_check_state = RETURN_CHECK_AWAITING_BRACE
else:
empty_return_check_state = RETURN_CHECK_INITIAL_STATE
if empty_return_check_state == RETURN_CHECK_AWAITING_BRACE:
empty_return_check_state = RETURN_CHECK_INITIAL_STATE
return True
return False
file_checks = [
{'regex': __regex_added_doc_rst,
'check': check_new_docs_index},
{'regex': __regex_added_doc_rst,
'check': check_doc_docs_automake}
]
checks = [
{'regex': None,
'match_name': lambda x: not line_length_ignore_list.search(x),
'check': lambda x: line_length_check(x)},
{'regex': None,
'match_name': lambda x: not leading_whitespace_ignore_list.search(x),
'check': lambda x: not leading_whitespace_is_spaces(x),
'print': lambda: print_warning("Line has non-spaces leading whitespace")},
{'regex': None, 'match_name': None,
'check': lambda x: trailing_whitespace_or_crlf(x),
'print': lambda: print_warning("Line has trailing whitespace")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': lambda x: not if_and_for_whitespace_checks(x),
'print': lambda: print_error("Improper whitespace around control block")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': lambda x: not if_and_for_end_with_bracket_check(x),
'print': lambda: print_error("Inappropriate bracing around statement")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': lambda x: pointer_whitespace_check(x),
'print':
lambda: print_error("Inappropriate spacing in pointer declaration")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'check': lambda x: nonascii_character_check(x),
'print':
lambda: print_error("Inappropriate non-ascii characters detected.")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': lambda x: cast_whitespace_check(x),
'print':
lambda: print_error("Inappropriate spacing around cast")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': lambda x: trailing_operator(x),
'print':
lambda: print_error("Line has '?' or ':' operator at end of line")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: has_comment(x),
'check': lambda x: has_xxx_mark(x),
'print': lambda: print_warning("Comment with 'xxx' marker")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': lambda x: has_c99_comment(x),
'print': lambda: print_error("C99 style comment")},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: has_comment(x),
'check': lambda x: check_spelling(x, True)},
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'check': lambda x: empty_return_with_brace(x),
'interim_line': True,
'print':
lambda: print_warning("Empty return followed by brace, consider omitting")
},
{'regex': r'(\.at|\.sh)$', 'match_name': None,
'check': lambda x: has_efgrep(x),
'print':
lambda: print_error("grep -E/-F should be used instead of egrep/fgrep")},
]
def regex_function_factory(func_name):
regex = re.compile(r'\b%s\([^)]*\)' % func_name)
return lambda x: regex.search(x) is not None
def regex_error_factory(description):
return lambda: print_error(description)
def regex_warn_factory(description):
return lambda: print_warning(description)
std_functions = [
('malloc', 'Use xmalloc() in place of malloc()'),
('calloc', 'Use xcalloc() in place of calloc()'),
('realloc', 'Use xrealloc() in place of realloc()'),
('strdup', 'Use xstrdup() in place of strdup()'),
('asprintf', 'Use xasprintf() in place of asprintf()'),
('vasprintf', 'Use xvasprintf() in place of vasprintf()'),
('strcpy', 'Use ovs_strlcpy() in place of strcpy()'),
('strlcpy', 'Use ovs_strlcpy() in place of strlcpy()'),
('strncpy', 'Use ovs_strzcpy() in place of strncpy()'),
('strerror', 'Use ovs_strerror() in place of strerror()'),
('sleep', 'Use xsleep() in place of sleep()'),
('abort', 'Use ovs_abort() in place of abort()'),
('assert', 'Use ovs_assert() in place of assert()'),
('error', 'Use ovs_error() in place of error()'),
]
checks += [
{'regex': r'(\.c|\.h)(\.in)?$',
'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': regex_function_factory(function_name),
'print': regex_error_factory(description)}
for (function_name, description) in std_functions]
easy_to_misuse_api = [
('ovsrcu_barrier',
'lib/ovs-rcu.c',
'Are you sure you need to use ovsrcu_barrier(), '
'in most cases ovsrcu_synchronize() will be fine?'),
]
checks += [
{'regex': r'(\.c)(\.in)?$',
'match_name': lambda x: x != location,
'prereq': lambda x: not is_comment_line(x),
'check': regex_function_factory(function_name),
'print': regex_warn_factory(description)}
for (function_name, location, description) in easy_to_misuse_api]
def regex_operator_factory(operator):
regex = re.compile(r'^[^#][^"\']*[^ "]%s[^ "\'][^"]*' % operator)
return lambda x: regex.search(filter_comments(x)) is not None
infix_operators = \
[re.escape(op) for op in ['%', '<<', '>>', '<=', '>=', '==', '!=',
'^', '|', '&&', '||', '?:', '=', '+=', '-=', '*=', '/=', '%=',
'&=', '^=', '|=', '<<=', '>>=']] \
+ [r'[^<" ]<[^=" ]',
r'[^\->" ]>[^=" ]',
r'[^ !()/"]\*[^/]',
r'[^ !&()"]&',
r'[^" +(]\+[^"+;]',
r'[^" \-(]\-[^"\->;]',
r'[^" <>=!^|+\-*/%&]=[^"=]',
r'[^* ]/[^* ]']
checks += [
{'regex': r'(\.c|\.h)(\.in)?$', 'match_name': None,
'prereq': lambda x: not is_comment_line(x),
'check': regex_operator_factory(operator),
'print': lambda: print_warning("Line lacks whitespace around operator")}
for operator in infix_operators]
def get_file_type_checks(filename):
"""Returns the list of checks for a file based on matching the filename
against regex."""
global checks
checkList = []
for check in checks:
regex_check = True
match_check = True
if check['regex'] is None and check['match_name'] is None:
checkList.append(check)
continue
if check['regex'] is not None and \
re.compile(check['regex']).search(filename) is None:
regex_check = False
if check['match_name'] is not None and \
not check['match_name'](filename):
match_check = False
if regex_check and match_check:
checkList.append(check)
return checkList
def run_checks(current_file, line, lineno):
"""Runs the various checks for the particular line. This will take
filename into account."""
global checking_file, total_line
print_line = False
for check in get_file_type_checks(current_file):
if 'prereq' in check and not check['prereq'](line):
continue
if check['check'](line):
if 'print' in check:
check['print']()
print_line = True
if print_line:
if checking_file:
print("%s:%d:" % (current_file, lineno))
else:
print("#%d FILE: %s:%d:" % (total_line, current_file, lineno))
print("%s\n" % line)
def interim_line_check(current_file, line, lineno):
"""Runs the various checks for the particular interim line. This will
take filename into account, and will check for the 'interim_line'
key before running the check."""
global checking_file, total_line
print_line = False
for check in get_file_type_checks(current_file):
if 'prereq' in check and not check['prereq'](line):
continue
if 'interim_line' in check and check['interim_line']:
if check['check'](line):
if 'print' in check:
check['print']()
print_line = True
if print_line:
if checking_file:
print("%s:%d:" % (current_file, lineno))
else:
print("#%d FILE: %s:%d:" % (total_line, current_file, lineno))
print("%s\n" % line)
def run_file_checks(text):
"""Runs the various checks for the text."""
for check in file_checks:
if check['regex'].search(text) is not None:
check['check'](text)
def ovs_checkpatch_parse(text, filename, author=None, committer=None):
global print_file_name, total_line, checking_file, \
empty_return_check_state
PARSE_STATE_HEADING = 0
PARSE_STATE_DIFF_HEADER = 1
PARSE_STATE_CHANGE_BODY = 2
lineno = 0
signatures = []
co_authors = []
parse = 0
current_file = filename if checking_file else ''
previous_file = ''
seppatch = re.compile(r'^---([\w]*| \S+)$')
hunks = re.compile(r'^(---|\+\+\+) (\S+)')
hunk_differences = re.compile(
r'^@@ ([0-9-+]+),([0-9-+]+) ([0-9-+]+),([0-9-+]+) @@')
is_author = re.compile(r'^(Author|From): (.*)$', re.I | re.M | re.S)
is_committer = re.compile(r'^(Commit: )(.*)$', re.I | re.M | re.S)
is_signature = re.compile(r'^(Signed-off-by: )(.*)$',
re.I | re.M | re.S)
is_co_author = re.compile(r'^(Co-authored-by: )(.*)$',
re.I | re.M | re.S)
is_gerrit_change_id = re.compile(r'(\s*(change-id: )(.*))$',
re.I | re.M | re.S)
is_fixes = re.compile(r'(\s*(Fixes:)(.*))$', re.I | re.M | re.S)
is_fixes_exact = re.compile(r'^Fixes: [0-9a-f]{12} \(".*"\)$')
tags_typos = {
r'^Acked by:': 'Acked-by:',
r'^Reported at:': 'Reported-at:',
r'^Reported by:': 'Reported-by:',
r'^Requested by:': 'Requested-by:',
r'^Reviewed by:': 'Reviewed-by:',
r'^Submitted at:': 'Submitted-at:',
r'^Suggested by:': 'Suggested-by:',
}
reset_counters()
for line in text.split("\n"):
if current_file != previous_file:
previous_file = current_file
lineno = lineno + 1
total_line = total_line + 1
if line == "\f":
# Form feed
continue
if len(line) <= 0:
continue
if checking_file:
parse = PARSE_STATE_CHANGE_BODY
if parse == PARSE_STATE_DIFF_HEADER:
match = hunks.match(line)
if match:
parse = PARSE_STATE_CHANGE_BODY
current_file = match.group(2)[2:]
print_file_name = current_file
continue
elif parse == PARSE_STATE_HEADING:
if seppatch.match(line):
parse = PARSE_STATE_DIFF_HEADER
if not skip_signoff_check:
# Check that the patch has an author, that the
# author is not among the co-authors, and that the
# co-authors are unique.
if not author:
print_error("Patch lacks author.")
continue
if " via " in author or "@openvswitch.org" in author:
print_error("Author should not be mailing list.")
continue
if author in co_authors:
print_error("Author should not be also be co-author.")
continue
if len(set(co_authors)) != len(co_authors):
print_error("Duplicate co-author.")
# Check that the author, all co-authors, and the
# committer (if any) signed off.
if author not in signatures:
print_error("Author %s needs to sign off." % author)
for ca in co_authors:
if ca not in signatures:
print_error("Co-author %s needs to sign off." % ca)
break
if (committer
and author != committer
and committer not in signatures):
print_error("Committer %s needs to sign off."
% committer)
# Check for signatures that we do not expect.
# This is only a warning because there can be,
# rarely, a signature chain.
#
# If we don't have a known committer, and there is
# a single extra sign-off, then do not warn
# because that extra sign-off is probably the
# committer.
extra_sigs = [x for x in signatures
if x not in co_authors
and x != author
and x != committer]
if len(extra_sigs) > 1 or (committer and extra_sigs):
print_warning("Unexpected sign-offs from developers "
"who are not authors or co-authors or "
"committers: %s"
% ", ".join(extra_sigs))
elif is_committer.match(line):
committer = is_committer.match(line).group(2)
elif is_author.match(line):
author = is_author.match(line).group(2)
elif is_signature.match(line):
m = is_signature.match(line)
signatures.append(m.group(2))
elif is_co_author.match(line):
m = is_co_author.match(line)
co_authors.append(m.group(2))
elif (is_gerrit_change_id.match(line) and
not skip_gerrit_change_id_check):
print_error(
"Remove Gerrit Change-Id's before submitting upstream.")
print("%d: %s\n" % (lineno, line))
elif is_fixes.match(line) and not is_fixes_exact.match(line):
print_error('"Fixes" tag is malformed.\n'
'Use the following format:\n'
' git log -1 '
'--pretty=format:"Fixes: %h (\\\"%s\\\")" '
'--abbrev=12 COMMIT_REF\n')
print("%d: %s\n" % (lineno, line))
elif spellcheck:
check_spelling(line, False)
for typo, correct in tags_typos.items():
m = re.match(typo, line, re.I)
if m:
print_error("%s tag is malformed." % (correct[:-1]))
print("%d: %s\n" % (lineno, line))
elif parse == PARSE_STATE_CHANGE_BODY:
newfile = hunks.match(line)
if newfile:
current_file = newfile.group(2)[2:]
print_file_name = current_file
continue
reset_line_number = hunk_differences.match(line)
if reset_line_number:
empty_return_check_state = RETURN_CHECK_INITIAL_STATE
lineno = int(reset_line_number.group(3))
if lineno < 0:
lineno = -1 * lineno
lineno -= 1
if is_subtracted_line(line):
lineno -= 1
continue
cmp_line = added_line(line)
if not is_added_line(line):
interim_line_check(current_file, cmp_line, lineno)
continue
# Skip files which have /datapath in them, since they are
# linux or windows coding standards
if current_file.startswith('datapath'):
continue
if current_file.startswith('include/linux'):
continue
# "sparse" includes could be copy-pasted from different sources
# like DPDK or Linux and could contain workarounds not suitable
# for a common style.
if current_file.startswith('include/sparse'):
continue
if current_file.startswith('utilities/bugtool'):
continue
run_checks(current_file, cmp_line, lineno)
run_file_checks(text)
if __errors or __warnings:
return EXIT_FAILURE
return 0
def usage():
print("""\
Open vSwitch checkpatch.py
Checks a patch for trivial mistakes.
usage:
%s [options] [PATCH1 [PATCH2 ...] | -f SOURCE1 [SOURCE2 ...] | -1 | -2 | ...]
Input options:
-f|--check-file Arguments are source files, not patches.
-1, -2, ... Check recent commits in this repo.
Check options:
-h|--help This help message
-b|--skip-block-whitespace Skips the if/while/for whitespace tests
-l|--skip-leading-whitespace Skips the leading whitespace test
-q|--quiet Only print error and warning information
-s|--skip-signoff-lines Tolerate missing Signed-off-by line
-S|--spellcheck Check C comments and commit-message for possible
spelling mistakes
-t|--skip-trailing-whitespace Skips the trailing whitespace test
--skip-gerrit-change-id Skips the gerrit change id test"""
% sys.argv[0])
def ovs_checkpatch_print_result():
global quiet, __warnings, __errors, total_line
if __errors or __warnings:
print("Lines checked: %d, Warnings: %d, Errors: %d\n" %
(total_line, __warnings, __errors))
elif not quiet:
print("Lines checked: %d, no obvious problems found\n" % (total_line))
def ovs_checkpatch_file(filename):
try:
mail = email.message_from_file(open(filename, 'r', encoding='utf8'))
except:
print_error("Unable to parse file '%s'. Is it a patch?" % filename)
return -1
for part in mail.walk():
if part.get_content_maintype() == 'multipart':
continue
result = ovs_checkpatch_parse(part.get_payload(decode=False), filename,
mail.get('Author', mail['From']),
mail['Commit'])
ovs_checkpatch_print_result()
return result
def partition(pred, iterable):
"""Returns [[trues], [falses]], where [trues] is the items in
'iterable' that satisfy 'pred' and [falses] is all the rest."""
trues = []
falses = []
for item in iterable:
if pred(item):
trues.append(item)
else:
falses.append(item)
return trues, falses
if __name__ == '__main__':
try:
numeric_options, args = partition(lambda s: re.match('-[0-9]+$', s),
sys.argv[1:])
n_patches = int(numeric_options[-1][1:]) if numeric_options else 0
optlist, args = getopt.getopt(args, 'bhlstfSq',
["check-file",
"help",
"skip-block-whitespace",
"skip-leading-whitespace",
"skip-signoff-lines",
"skip-trailing-whitespace",
"skip-gerrit-change-id",
"spellcheck",
"quiet"])
except:
print("Unknown option encountered. Please rerun with -h for help.")
sys.exit(EXIT_FAILURE)
for o, a in optlist:
if o in ("-h", "--help"):
usage()
sys.exit(0)
elif o in ("-b", "--skip-block-whitespace"):
skip_block_whitespace_check = True
elif o in ("-l", "--skip-leading-whitespace"):
skip_leading_whitespace_check = True
elif o in ("-s", "--skip-signoff-lines"):
skip_signoff_check = True
elif o in ("-t", "--skip-trailing-whitespace"):
skip_trailing_whitespace_check = True
elif o in ("--skip-gerrit-change-id"):
skip_gerrit_change_id_check = True
elif o in ("-f", "--check-file"):
checking_file = True
elif o in ("-S", "--spellcheck"):
if not open_spell_check_dict():
print("WARNING: The enchant library isn't available.")
print(" Please install python enchant.")
else:
spellcheck = True
elif o in ("-q", "--quiet"):
quiet = True
else:
print("Unknown option '%s'" % o)
sys.exit(EXIT_FAILURE)
if sys.stdout.isatty():
colors = True
if n_patches:
status = 0
git_log = 'git log --no-color --no-merges --pretty=format:"%H %s" '
with os.popen(git_log + '-%d' % n_patches, 'r') as f:
commits = f.read().split("\n")
for i in reversed(range(0, n_patches)):
revision, name = commits[i].split(" ", 1)
f = os.popen('''git format-patch -1 --stdout --pretty=format:"\
Author: %an <%ae>
Commit: %cn <%ce>
Subject: %s
%b" ''' + revision, 'r')
patch = f.read()
f.close()
if not quiet:
print('== Checking %s ("%s") ==' % (revision[0:12], name))
result = ovs_checkpatch_parse(patch, revision)
ovs_checkpatch_print_result()
if result:
status = EXIT_FAILURE
sys.exit(status)
if not args:
if sys.stdin.isatty():
usage()
sys.exit(EXIT_FAILURE)
result = ovs_checkpatch_parse(sys.stdin.read(), '-')
ovs_checkpatch_print_result()
sys.exit(result)
status = 0
for filename in args:
if not quiet:
print('== Checking "%s" ==' % filename)
result = ovs_checkpatch_file(filename)
if result:
status = EXIT_FAILURE
sys.exit(status)