mirror of
https://github.com/PyCQA/flake8.git
synced 2026-03-29 10:36:53 +00:00
extend black formatting to tests as well
This commit is contained in:
parent
a7174759e9
commit
af1668bf04
45 changed files with 1644 additions and 1307 deletions
|
|
@ -1,14 +1,12 @@
|
|||
exclude: ^tests/fixtures/example-code/
|
||||
exclude: ^tests/fixtures/
|
||||
repos:
|
||||
- repo: https://github.com/pre-commit/pre-commit-hooks
|
||||
rev: v3.4.0
|
||||
hooks:
|
||||
- id: check-yaml
|
||||
- id: debug-statements
|
||||
exclude: ^tests/fixtures/example-code/invalid-syntax.py$
|
||||
- id: end-of-file-fixer
|
||||
- id: trailing-whitespace
|
||||
exclude: ^tests/fixtures/diffs/
|
||||
- repo: https://github.com/asottile/reorder_python_imports
|
||||
rev: v2.4.0
|
||||
hooks:
|
||||
|
|
@ -18,8 +16,7 @@ repos:
|
|||
rev: 20.8b1
|
||||
hooks:
|
||||
- id: black
|
||||
args: [--line-length=78]
|
||||
files: ^src/
|
||||
args: [--line-length=79]
|
||||
- repo: https://github.com/asottile/pyupgrade
|
||||
rev: v2.12.0
|
||||
hooks:
|
||||
|
|
@ -29,4 +26,4 @@ repos:
|
|||
rev: v0.812
|
||||
hooks:
|
||||
- id: mypy
|
||||
exclude: ^(docs/|example-plugin/|tests/fixtures)
|
||||
exclude: ^(docs/|example-plugin/)
|
||||
|
|
|
|||
|
|
@ -16,45 +16,45 @@ import sys
|
|||
# If extensions (or modules to document with autodoc) are in another directory,
|
||||
# add these directories to sys.path here. If the directory is relative to the
|
||||
# documentation root, use os.path.abspath to make it absolute, like shown here.
|
||||
#sys.path.insert(0, os.path.abspath('.'))
|
||||
# sys.path.insert(0, os.path.abspath('.'))
|
||||
|
||||
# -- General configuration ------------------------------------------------
|
||||
|
||||
# If your documentation needs a minimal Sphinx version, state it here.
|
||||
needs_sphinx = '1.3'
|
||||
needs_sphinx = "1.3"
|
||||
|
||||
# Add any Sphinx extension module names here, as strings. They can be
|
||||
# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
|
||||
# ones.
|
||||
extensions = [
|
||||
'sphinx.ext.autodoc',
|
||||
'sphinx.ext.doctest',
|
||||
'sphinx.ext.extlinks',
|
||||
'sphinx.ext.intersphinx',
|
||||
'sphinx.ext.todo',
|
||||
'sphinx.ext.coverage',
|
||||
'sphinx.ext.viewcode',
|
||||
'sphinx-prompt',
|
||||
"sphinx.ext.autodoc",
|
||||
"sphinx.ext.doctest",
|
||||
"sphinx.ext.extlinks",
|
||||
"sphinx.ext.intersphinx",
|
||||
"sphinx.ext.todo",
|
||||
"sphinx.ext.coverage",
|
||||
"sphinx.ext.viewcode",
|
||||
"sphinx-prompt",
|
||||
]
|
||||
|
||||
# Add any paths that contain templates here, relative to this directory.
|
||||
templates_path = ['_templates']
|
||||
templates_path = ["_templates"]
|
||||
|
||||
# The suffix(es) of source filenames.
|
||||
# You can specify multiple suffix as a list of string:
|
||||
# source_suffix = ['.rst', '.md']
|
||||
source_suffix = '.rst'
|
||||
source_suffix = ".rst"
|
||||
|
||||
# The encoding of source files.
|
||||
#source_encoding = 'utf-8-sig'
|
||||
# source_encoding = 'utf-8-sig'
|
||||
|
||||
# The master toctree document.
|
||||
master_doc = 'index'
|
||||
master_doc = "index"
|
||||
|
||||
# General information about the project.
|
||||
project = 'flake8'
|
||||
copyright = '2016, Ian Stapleton Cordasco'
|
||||
author = 'Ian Stapleton Cordasco'
|
||||
project = "flake8"
|
||||
copyright = "2016, Ian Stapleton Cordasco"
|
||||
author = "Ian Stapleton Cordasco"
|
||||
|
||||
import flake8
|
||||
|
||||
|
|
@ -80,9 +80,9 @@ language = None
|
|||
|
||||
# There are two options for replacing |today|: either, you set today to some
|
||||
# non-false value, then it is used:
|
||||
#today = ''
|
||||
# today = ''
|
||||
# Else, today_fmt is used as the format for a strftime call.
|
||||
#today_fmt = '%B %d, %Y'
|
||||
# today_fmt = '%B %d, %Y'
|
||||
|
||||
# List of patterns, relative to source directory, that match files and
|
||||
# directories to ignore when looking for source files.
|
||||
|
|
@ -90,27 +90,27 @@ exclude_patterns = []
|
|||
|
||||
# The reST default role (used for this markup: `text`) to use for all
|
||||
# documents.
|
||||
#default_role = None
|
||||
# default_role = None
|
||||
|
||||
# If true, '()' will be appended to :func: etc. cross-reference text.
|
||||
#add_function_parentheses = True
|
||||
# add_function_parentheses = True
|
||||
|
||||
# If true, the current module name will be prepended to all description
|
||||
# unit titles (such as .. function::).
|
||||
#add_module_names = True
|
||||
# add_module_names = True
|
||||
|
||||
# If true, sectionauthor and moduleauthor directives will be shown in the
|
||||
# output. They are ignored by default.
|
||||
#show_authors = False
|
||||
# show_authors = False
|
||||
|
||||
# The name of the Pygments (syntax highlighting) style to use.
|
||||
pygments_style = 'sphinx'
|
||||
pygments_style = "sphinx"
|
||||
|
||||
# A list of ignored prefixes for module index sorting.
|
||||
#modindex_common_prefix = []
|
||||
# modindex_common_prefix = []
|
||||
|
||||
# If true, keep warnings as "system message" paragraphs in the built documents.
|
||||
#keep_warnings = False
|
||||
# keep_warnings = False
|
||||
|
||||
# If true, `todo` and `todoList` produce output, else they produce nothing.
|
||||
todo_include_todos = True
|
||||
|
|
@ -120,31 +120,31 @@ todo_include_todos = True
|
|||
|
||||
# The theme to use for HTML and HTML Help pages. See the documentation for
|
||||
# a list of builtin themes.
|
||||
html_theme = 'sphinx_rtd_theme'
|
||||
html_theme = "sphinx_rtd_theme"
|
||||
|
||||
# Theme options are theme-specific and customize the look and feel of a theme
|
||||
# further. For a list of options available for each theme, see the
|
||||
# documentation.
|
||||
#html_theme_options = {}
|
||||
# html_theme_options = {}
|
||||
|
||||
# Add any paths that contain custom themes here, relative to this directory.
|
||||
#html_theme_path = []
|
||||
# html_theme_path = []
|
||||
|
||||
# The name for this set of Sphinx documents. If None, it defaults to
|
||||
# "<project> v<release> documentation".
|
||||
#html_title = None
|
||||
# html_title = None
|
||||
|
||||
# A shorter title for the navigation bar. Default is the same as html_title.
|
||||
#html_short_title = None
|
||||
# html_short_title = None
|
||||
|
||||
# The name of an image file (relative to this directory) to place at the top
|
||||
# of the sidebar.
|
||||
#html_logo = None
|
||||
# html_logo = None
|
||||
|
||||
# The name of an image file (within the static path) to use as favicon of the
|
||||
# docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32
|
||||
# pixels large.
|
||||
#html_favicon = None
|
||||
# html_favicon = None
|
||||
|
||||
# Add any paths that contain custom static files (such as style sheets) here,
|
||||
# relative to this directory. They are copied after the builtin static files,
|
||||
|
|
@ -154,109 +154,111 @@ html_theme = 'sphinx_rtd_theme'
|
|||
# Add any extra paths that contain custom files (such as robots.txt or
|
||||
# .htaccess) here, relative to this directory. These files are copied
|
||||
# directly to the root of the documentation.
|
||||
#html_extra_path = []
|
||||
# html_extra_path = []
|
||||
|
||||
# If not '', a 'Last updated on:' timestamp is inserted at every page bottom,
|
||||
# using the given strftime format.
|
||||
#html_last_updated_fmt = '%b %d, %Y'
|
||||
# html_last_updated_fmt = '%b %d, %Y'
|
||||
|
||||
# If true, SmartyPants will be used to convert quotes and dashes to
|
||||
# typographically correct entities.
|
||||
#html_use_smartypants = True
|
||||
# html_use_smartypants = True
|
||||
|
||||
# Custom sidebar templates, maps document names to template names.
|
||||
#html_sidebars = {}
|
||||
# html_sidebars = {}
|
||||
|
||||
# Additional templates that should be rendered to pages, maps page names to
|
||||
# template names.
|
||||
#html_additional_pages = {}
|
||||
# html_additional_pages = {}
|
||||
|
||||
# If false, no module index is generated.
|
||||
#html_domain_indices = True
|
||||
# html_domain_indices = True
|
||||
|
||||
# If false, no index is generated.
|
||||
#html_use_index = True
|
||||
# html_use_index = True
|
||||
|
||||
# If true, the index is split into individual pages for each letter.
|
||||
#html_split_index = False
|
||||
# html_split_index = False
|
||||
|
||||
# If true, links to the reST sources are added to the pages.
|
||||
#html_show_sourcelink = True
|
||||
# html_show_sourcelink = True
|
||||
|
||||
# If true, "Created using Sphinx" is shown in the HTML footer. Default is True.
|
||||
#html_show_sphinx = True
|
||||
# html_show_sphinx = True
|
||||
|
||||
# If true, "(C) Copyright ..." is shown in the HTML footer. Default is True.
|
||||
#html_show_copyright = True
|
||||
# html_show_copyright = True
|
||||
|
||||
# If true, an OpenSearch description file will be output, and all pages will
|
||||
# contain a <link> tag referring to it. The value of this option must be the
|
||||
# base URL from which the finished HTML is served.
|
||||
#html_use_opensearch = ''
|
||||
# html_use_opensearch = ''
|
||||
|
||||
# This is the file name suffix for HTML files (e.g. ".xhtml").
|
||||
#html_file_suffix = None
|
||||
# html_file_suffix = None
|
||||
|
||||
# Language to be used for generating the HTML full-text search index.
|
||||
# Sphinx supports the following languages:
|
||||
# 'da', 'de', 'en', 'es', 'fi', 'fr', 'hu', 'it', 'ja'
|
||||
# 'nl', 'no', 'pt', 'ro', 'ru', 'sv', 'tr'
|
||||
#html_search_language = 'en'
|
||||
# html_search_language = 'en'
|
||||
|
||||
# A dictionary with options for the search language support, empty by default.
|
||||
# Now only 'ja' uses this config value
|
||||
#html_search_options = {'type': 'default'}
|
||||
# html_search_options = {'type': 'default'}
|
||||
|
||||
# The name of a javascript file (relative to the configuration directory) that
|
||||
# implements a search results scorer. If empty, the default will be used.
|
||||
#html_search_scorer = 'scorer.js'
|
||||
# html_search_scorer = 'scorer.js'
|
||||
|
||||
# Output file base name for HTML help builder.
|
||||
htmlhelp_basename = 'flake8doc'
|
||||
htmlhelp_basename = "flake8doc"
|
||||
|
||||
# -- Options for LaTeX output ---------------------------------------------
|
||||
|
||||
latex_elements = {
|
||||
# The paper size ('letterpaper' or 'a4paper').
|
||||
#'papersize': 'letterpaper',
|
||||
|
||||
# The font size ('10pt', '11pt' or '12pt').
|
||||
#'pointsize': '10pt',
|
||||
|
||||
# Additional stuff for the LaTeX preamble.
|
||||
#'preamble': '',
|
||||
|
||||
# Latex figure (float) alignment
|
||||
#'figure_align': 'htbp',
|
||||
# The paper size ('letterpaper' or 'a4paper').
|
||||
#'papersize': 'letterpaper',
|
||||
# The font size ('10pt', '11pt' or '12pt').
|
||||
#'pointsize': '10pt',
|
||||
# Additional stuff for the LaTeX preamble.
|
||||
#'preamble': '',
|
||||
# Latex figure (float) alignment
|
||||
#'figure_align': 'htbp',
|
||||
}
|
||||
|
||||
# Grouping the document tree into LaTeX files. List of tuples
|
||||
# (source start file, target name, title,
|
||||
# author, documentclass [howto, manual, or own class]).
|
||||
latex_documents = [
|
||||
(master_doc, 'flake8.tex', 'flake8 Documentation',
|
||||
'Ian Stapleton Cordasco', 'manual'),
|
||||
(
|
||||
master_doc,
|
||||
"flake8.tex",
|
||||
"flake8 Documentation",
|
||||
"Ian Stapleton Cordasco",
|
||||
"manual",
|
||||
),
|
||||
]
|
||||
|
||||
# The name of an image file (relative to this directory) to place at the top of
|
||||
# the title page.
|
||||
#latex_logo = None
|
||||
# latex_logo = None
|
||||
|
||||
# For "manual" documents, if this is true, then toplevel headings are parts,
|
||||
# not chapters.
|
||||
#latex_use_parts = False
|
||||
# latex_use_parts = False
|
||||
|
||||
# If true, show page references after internal links.
|
||||
#latex_show_pagerefs = False
|
||||
# latex_show_pagerefs = False
|
||||
|
||||
# If true, show URL addresses after external links.
|
||||
#latex_show_urls = False
|
||||
# latex_show_urls = False
|
||||
|
||||
# Documents to append as an appendix to all manuals.
|
||||
#latex_appendices = []
|
||||
# latex_appendices = []
|
||||
|
||||
# If false, no module index is generated.
|
||||
#latex_domain_indices = True
|
||||
# latex_domain_indices = True
|
||||
|
||||
|
||||
# -- Options for manual page output ---------------------------------------
|
||||
|
|
@ -264,12 +266,11 @@ latex_documents = [
|
|||
# One entry per manual page. List of tuples
|
||||
# (source start file, name, description, authors, manual section).
|
||||
man_pages = [
|
||||
('manpage', 'flake8', 'Flake8 Command Line Documentation',
|
||||
[author], 1)
|
||||
("manpage", "flake8", "Flake8 Command Line Documentation", [author], 1)
|
||||
]
|
||||
|
||||
# If true, show URL addresses after external links.
|
||||
#man_show_urls = False
|
||||
# man_show_urls = False
|
||||
|
||||
|
||||
# -- Options for Texinfo output -------------------------------------------
|
||||
|
|
@ -278,26 +279,32 @@ man_pages = [
|
|||
# (source start file, target name, title, author,
|
||||
# dir menu entry, description, category)
|
||||
texinfo_documents = [
|
||||
('index', 'Flake8', 'Flake8 Documentation', 'Tarek Ziade',
|
||||
'Flake8', 'Code checking using pycodestyle, pyflakes and mccabe',
|
||||
'Miscellaneous'),
|
||||
(
|
||||
"index",
|
||||
"Flake8",
|
||||
"Flake8 Documentation",
|
||||
"Tarek Ziade",
|
||||
"Flake8",
|
||||
"Code checking using pycodestyle, pyflakes and mccabe",
|
||||
"Miscellaneous",
|
||||
),
|
||||
]
|
||||
|
||||
# Documents to append as an appendix to all manuals.
|
||||
#texinfo_appendices = []
|
||||
# texinfo_appendices = []
|
||||
|
||||
# If false, no module index is generated.
|
||||
#texinfo_domain_indices = True
|
||||
# texinfo_domain_indices = True
|
||||
|
||||
# How to display URL addresses: 'footnote', 'no', or 'inline'.
|
||||
#texinfo_show_urls = 'footnote'
|
||||
# texinfo_show_urls = 'footnote'
|
||||
|
||||
# If true, do not generate a @detailmenu in the "Top" node's menu.
|
||||
#texinfo_no_detailmenu = False
|
||||
# texinfo_no_detailmenu = False
|
||||
|
||||
|
||||
# Example configuration for intersphinx: refer to the Python standard library.
|
||||
intersphinx_mapping = {'python': ('https://docs.python.org/3/', None)}
|
||||
intersphinx_mapping = {"python": ("https://docs.python.org/3/", None)}
|
||||
|
||||
extlinks = {
|
||||
"issue": ("https://github.com/pycqa/flake8/issues/%s", "#"),
|
||||
|
|
|
|||
|
|
@ -1,29 +1,29 @@
|
|||
import setuptools
|
||||
|
||||
setuptools.setup(
|
||||
name='flake8-example-plugin',
|
||||
license='MIT',
|
||||
version='1.0.0',
|
||||
description='Example plugin to Flake8',
|
||||
author='Ian Cordasco',
|
||||
author_email='graffatcolmingov@gmail.com',
|
||||
url='https://github.com/pycqa/flake8',
|
||||
package_dir={'': 'src/'},
|
||||
packages=['flake8_example_plugin'],
|
||||
name="flake8-example-plugin",
|
||||
license="MIT",
|
||||
version="1.0.0",
|
||||
description="Example plugin to Flake8",
|
||||
author="Ian Cordasco",
|
||||
author_email="graffatcolmingov@gmail.com",
|
||||
url="https://github.com/pycqa/flake8",
|
||||
package_dir={"": "src/"},
|
||||
packages=["flake8_example_plugin"],
|
||||
entry_points={
|
||||
'flake8.extension': [
|
||||
'X1 = flake8_example_plugin:ExampleOne',
|
||||
'X2 = flake8_example_plugin:ExampleTwo',
|
||||
"flake8.extension": [
|
||||
"X1 = flake8_example_plugin:ExampleOne",
|
||||
"X2 = flake8_example_plugin:ExampleTwo",
|
||||
],
|
||||
},
|
||||
classifiers=[
|
||||
'Framework :: Flake8',
|
||||
'License :: OSI Approved :: MIT License',
|
||||
'Programming Language :: Python',
|
||||
'Programming Language :: Python :: 3',
|
||||
'Programming Language :: Python :: 3.8',
|
||||
'Programming Language :: Python :: 3.9',
|
||||
'Topic :: Software Development :: Libraries :: Python Modules',
|
||||
'Topic :: Software Development :: Quality Assurance',
|
||||
"Framework :: Flake8",
|
||||
"License :: OSI Approved :: MIT License",
|
||||
"Programming Language :: Python",
|
||||
"Programming Language :: Python :: 3",
|
||||
"Programming Language :: Python :: 3.8",
|
||||
"Programming Language :: Python :: 3.9",
|
||||
"Topic :: Software Development :: Libraries :: Python Modules",
|
||||
"Topic :: Software Development :: Quality Assurance",
|
||||
],
|
||||
)
|
||||
|
|
|
|||
|
|
@ -3,6 +3,6 @@ from .off_by_default import ExampleTwo
|
|||
from .on_by_default import ExampleOne
|
||||
|
||||
__all__ = (
|
||||
'ExampleOne',
|
||||
'ExampleTwo',
|
||||
"ExampleOne",
|
||||
"ExampleTwo",
|
||||
)
|
||||
|
|
|
|||
|
|
@ -3,8 +3,9 @@
|
|||
|
||||
class ExampleTwo:
|
||||
"""Second Example Plugin."""
|
||||
name = 'off-by-default-example-plugin'
|
||||
version = '1.0.0'
|
||||
|
||||
name = "off-by-default-example-plugin"
|
||||
version = "1.0.0"
|
||||
|
||||
off_by_default = True
|
||||
|
||||
|
|
@ -13,5 +14,9 @@ class ExampleTwo:
|
|||
|
||||
def run(self):
|
||||
"""Do nothing."""
|
||||
yield (1, 0, 'X200 The off-by-default plugin was enabled',
|
||||
'OffByDefaultPlugin')
|
||||
yield (
|
||||
1,
|
||||
0,
|
||||
"X200 The off-by-default plugin was enabled",
|
||||
"OffByDefaultPlugin",
|
||||
)
|
||||
|
|
|
|||
|
|
@ -3,8 +3,9 @@
|
|||
|
||||
class ExampleOne:
|
||||
"""First Example Plugin."""
|
||||
name = 'on-by-default-example-plugin'
|
||||
version = '1.0.0'
|
||||
|
||||
name = "on-by-default-example-plugin"
|
||||
version = "1.0.0"
|
||||
|
||||
def __init__(self, tree):
|
||||
self.tree = tree
|
||||
|
|
|
|||
2
setup.py
2
setup.py
|
|
@ -4,6 +4,6 @@ import sys
|
|||
|
||||
import setuptools
|
||||
|
||||
sys.path.insert(0, os.path.join(os.path.dirname(__file__), 'src'))
|
||||
sys.path.insert(0, os.path.join(os.path.dirname(__file__), "src"))
|
||||
|
||||
setuptools.setup()
|
||||
|
|
|
|||
|
|
@ -17,9 +17,7 @@ LOG = logging.getLogger(__name__)
|
|||
LOG.addHandler(logging.NullHandler())
|
||||
|
||||
__version__ = "3.9.1"
|
||||
__version_info__ = tuple(
|
||||
int(i) for i in __version__.split(".") if i.isdigit()
|
||||
)
|
||||
__version_info__ = tuple(int(i) for i in __version__.split(".") if i.isdigit())
|
||||
|
||||
|
||||
# There is nothing lower than logging.DEBUG (10) in the logging library,
|
||||
|
|
|
|||
|
|
@ -241,9 +241,7 @@ class Manager:
|
|||
"""
|
||||
results_reported = results_found = 0
|
||||
for checker in self._all_checkers:
|
||||
results = sorted(
|
||||
checker.results, key=lambda tup: (tup[1], tup[2])
|
||||
)
|
||||
results = sorted(checker.results, key=lambda tup: (tup[1], tup[2]))
|
||||
filename = checker.display_name
|
||||
with self.style_guide.processing_file(filename):
|
||||
results_reported += self._handle_results(filename, results)
|
||||
|
|
|
|||
|
|
@ -39,9 +39,7 @@ class InvalidSyntax(Flake8Exception):
|
|||
def __init__(self, exception: Exception) -> None:
|
||||
"""Initialize our InvalidSyntax exception."""
|
||||
self.original_exception = exception
|
||||
self.error_message = (
|
||||
f"{type(exception).__name__}: {exception.args[0]}"
|
||||
)
|
||||
self.error_message = f"{type(exception).__name__}: {exception.args[0]}"
|
||||
self.error_code = "E902"
|
||||
self.line_number = 1
|
||||
self.column_number = 0
|
||||
|
|
|
|||
|
|
@ -369,6 +369,4 @@ def get_local_plugins(config_finder):
|
|||
return local_plugins
|
||||
|
||||
|
||||
LocalPlugins = collections.namedtuple(
|
||||
"LocalPlugins", "extension report paths"
|
||||
)
|
||||
LocalPlugins = collections.namedtuple("LocalPlugins", "extension report paths")
|
||||
|
|
|
|||
|
|
@ -472,8 +472,7 @@ class Checkers(PluginTypeManager):
|
|||
plugin.to_dictionary() for plugin in self.logical_line_plugins
|
||||
],
|
||||
"physical_line_plugins": [
|
||||
plugin.to_dictionary()
|
||||
for plugin in self.physical_line_plugins
|
||||
plugin.to_dictionary() for plugin in self.physical_line_plugins
|
||||
],
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -289,9 +289,7 @@ class FileProcessor:
|
|||
except (tokenize.TokenError, SyntaxError) as exc:
|
||||
raise exceptions.InvalidSyntax(exception=exc)
|
||||
|
||||
def _noqa_line_range(
|
||||
self, min_line: int, max_line: int
|
||||
) -> Dict[int, str]:
|
||||
def _noqa_line_range(self, min_line: int, max_line: int) -> Dict[int, str]:
|
||||
line_range = range(min_line, max_line + 1)
|
||||
joined = "".join(self.lines[min_line - 1 : max_line])
|
||||
return dict.fromkeys(line_range, joined)
|
||||
|
|
|
|||
|
|
@ -368,9 +368,7 @@ class StyleGuideManager:
|
|||
:rtype:
|
||||
:class:`~flake8.style_guide.StyleGuide`
|
||||
"""
|
||||
per_file = utils.parse_files_to_codes_mapping(
|
||||
options.per_file_ignores
|
||||
)
|
||||
per_file = utils.parse_files_to_codes_mapping(options.per_file_ignores)
|
||||
for filename, violations in per_file:
|
||||
yield self.default_style_guide.copy(
|
||||
filename=filename, extend_ignore_with=violations
|
||||
|
|
@ -579,11 +577,7 @@ class StyleGuide:
|
|||
)
|
||||
is_not_inline_ignored = error.is_inline_ignored(disable_noqa) is False
|
||||
is_included_in_diff = error.is_in(self._parsed_diff)
|
||||
if (
|
||||
error_is_selected
|
||||
and is_not_inline_ignored
|
||||
and is_included_in_diff
|
||||
):
|
||||
if error_is_selected and is_not_inline_ignored and is_included_in_diff:
|
||||
self.formatter.handle(error)
|
||||
self.stats.record(error)
|
||||
return 1
|
||||
|
|
|
|||
|
|
@ -271,13 +271,10 @@ def parse_unified_diff(diff: Optional[str] = None) -> Dict[str, Set[int]]:
|
|||
# comparing.
|
||||
if hunk_match:
|
||||
(row, number_of_rows) = [
|
||||
1 if not group else int(group)
|
||||
for group in hunk_match.groups()
|
||||
1 if not group else int(group) for group in hunk_match.groups()
|
||||
]
|
||||
assert current_path is not None
|
||||
parsed_paths[current_path].update(
|
||||
range(row, row + number_of_rows)
|
||||
)
|
||||
parsed_paths[current_path].update(range(row, row + number_of_rows))
|
||||
|
||||
# We have now parsed our diff into a dictionary that looks like:
|
||||
# {'file.py': set(range(10, 16), range(18, 20)), ...}
|
||||
|
|
|
|||
|
|
@ -3,4 +3,4 @@ import sys
|
|||
|
||||
import flake8
|
||||
|
||||
flake8.configure_logging(2, 'test-logs-%s.%s.log' % sys.version_info[0:2])
|
||||
flake8.configure_logging(2, "test-logs-%s.%s.log" % sys.version_info[0:2])
|
||||
|
|
|
|||
|
|
@ -4,8 +4,8 @@
|
|||
class ExtensionTestPlugin2:
|
||||
"""Extension test plugin in its own directory."""
|
||||
|
||||
name = 'ExtensionTestPlugin2'
|
||||
version = '1.0.0'
|
||||
name = "ExtensionTestPlugin2"
|
||||
version = "1.0.0"
|
||||
|
||||
def __init__(self, tree):
|
||||
"""Construct an instance of test plugin."""
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ from flake8.options import aggregator
|
|||
from flake8.options import config
|
||||
from flake8.options import manager
|
||||
|
||||
CLI_SPECIFIED_CONFIG = 'tests/fixtures/config_files/cli-specified.ini'
|
||||
CLI_SPECIFIED_CONFIG = "tests/fixtures/config_files/cli-specified.ini"
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
|
|
@ -18,8 +18,8 @@ def optmanager():
|
|||
prelim_parser = argparse.ArgumentParser(add_help=False)
|
||||
options.register_preliminary_options(prelim_parser)
|
||||
option_manager = manager.OptionManager(
|
||||
prog='flake8',
|
||||
version='3.0.0',
|
||||
prog="flake8",
|
||||
version="3.0.0",
|
||||
parents=[prelim_parser],
|
||||
)
|
||||
options.register_default_options(option_manager)
|
||||
|
|
@ -28,31 +28,50 @@ def optmanager():
|
|||
|
||||
def test_aggregate_options_with_config(optmanager):
|
||||
"""Verify we aggregate options and config values appropriately."""
|
||||
arguments = ['flake8', '--select',
|
||||
'E11,E34,E402,W,F', '--exclude', 'tests/*']
|
||||
arguments = [
|
||||
"flake8",
|
||||
"--select",
|
||||
"E11,E34,E402,W,F",
|
||||
"--exclude",
|
||||
"tests/*",
|
||||
]
|
||||
config_finder = config.ConfigFileFinder(
|
||||
'flake8',
|
||||
config_file=CLI_SPECIFIED_CONFIG)
|
||||
"flake8", config_file=CLI_SPECIFIED_CONFIG
|
||||
)
|
||||
options, args = aggregator.aggregate_options(
|
||||
optmanager, config_finder, arguments)
|
||||
optmanager, config_finder, arguments
|
||||
)
|
||||
|
||||
assert options.select == ['E11', 'E34', 'E402', 'W', 'F']
|
||||
assert options.ignore == ['E123', 'W234', 'E111']
|
||||
assert options.exclude == [os.path.abspath('tests/*')]
|
||||
assert options.select == ["E11", "E34", "E402", "W", "F"]
|
||||
assert options.ignore == ["E123", "W234", "E111"]
|
||||
assert options.exclude == [os.path.abspath("tests/*")]
|
||||
|
||||
|
||||
def test_aggregate_options_when_isolated(optmanager):
|
||||
"""Verify we aggregate options and config values appropriately."""
|
||||
arguments = ['flake8', '--select', 'E11,E34,E402,W,F',
|
||||
'--exclude', 'tests/*']
|
||||
config_finder = config.ConfigFileFinder(
|
||||
'flake8', ignore_config_files=True)
|
||||
optmanager.extend_default_ignore(['E8'])
|
||||
options, args = aggregator.aggregate_options(
|
||||
optmanager, config_finder, arguments)
|
||||
|
||||
assert options.select == ['E11', 'E34', 'E402', 'W', 'F']
|
||||
assert sorted(options.ignore) == [
|
||||
'E121', 'E123', 'E126', 'E226', 'E24', 'E704', 'E8', 'W503', 'W504',
|
||||
arguments = [
|
||||
"flake8",
|
||||
"--select",
|
||||
"E11,E34,E402,W,F",
|
||||
"--exclude",
|
||||
"tests/*",
|
||||
]
|
||||
assert options.exclude == [os.path.abspath('tests/*')]
|
||||
config_finder = config.ConfigFileFinder("flake8", ignore_config_files=True)
|
||||
optmanager.extend_default_ignore(["E8"])
|
||||
options, args = aggregator.aggregate_options(
|
||||
optmanager, config_finder, arguments
|
||||
)
|
||||
|
||||
assert options.select == ["E11", "E34", "E402", "W", "F"]
|
||||
assert sorted(options.ignore) == [
|
||||
"E121",
|
||||
"E123",
|
||||
"E126",
|
||||
"E226",
|
||||
"E24",
|
||||
"E704",
|
||||
"E8",
|
||||
"W503",
|
||||
"W504",
|
||||
]
|
||||
assert options.exclude == [os.path.abspath("tests/*")]
|
||||
|
|
|
|||
|
|
@ -5,8 +5,8 @@ from flake8.api import legacy
|
|||
def test_legacy_api(tmpdir):
|
||||
"""A basic end-to-end test for the legacy api reporting errors."""
|
||||
with tmpdir.as_cwd():
|
||||
t_py = tmpdir.join('t.py')
|
||||
t_py.write('import os # unused import\n')
|
||||
t_py = tmpdir.join("t.py")
|
||||
t_py.write("import os # unused import\n")
|
||||
|
||||
style_guide = legacy.get_style_guide()
|
||||
report = style_guide.check_files([t_py.strpath])
|
||||
|
|
|
|||
|
|
@ -10,13 +10,13 @@ from flake8.processor import FileProcessor
|
|||
|
||||
PHYSICAL_LINE = "# Physical line content"
|
||||
|
||||
EXPECTED_REPORT = (1, 1, 'T000 Expected Message')
|
||||
EXPECTED_REPORT_PHYSICAL_LINE = (1, 'T000 Expected Message')
|
||||
EXPECTED_REPORT = (1, 1, "T000 Expected Message")
|
||||
EXPECTED_REPORT_PHYSICAL_LINE = (1, "T000 Expected Message")
|
||||
EXPECTED_RESULT_PHYSICAL_LINE = (
|
||||
'T000',
|
||||
"T000",
|
||||
0,
|
||||
1,
|
||||
'Expected Message',
|
||||
"Expected Message",
|
||||
None,
|
||||
)
|
||||
|
||||
|
|
@ -24,8 +24,8 @@ EXPECTED_RESULT_PHYSICAL_LINE = (
|
|||
class PluginClass:
|
||||
"""Simple file plugin class yielding the expected report."""
|
||||
|
||||
name = 'test'
|
||||
version = '1.0.0'
|
||||
name = "test"
|
||||
version = "1.0.0"
|
||||
|
||||
def __init__(self, tree):
|
||||
"""Construct a dummy object to provide mandatory parameter."""
|
||||
|
|
@ -33,26 +33,26 @@ class PluginClass:
|
|||
|
||||
def run(self):
|
||||
"""Run class yielding one element containing the expected report."""
|
||||
yield EXPECTED_REPORT + (type(self), )
|
||||
yield EXPECTED_REPORT + (type(self),)
|
||||
|
||||
|
||||
def plugin_func(func):
|
||||
"""Decorate file plugins which are implemented as functions."""
|
||||
func.name = 'test'
|
||||
func.version = '1.0.0'
|
||||
func.name = "test"
|
||||
func.version = "1.0.0"
|
||||
return func
|
||||
|
||||
|
||||
@plugin_func
|
||||
def plugin_func_gen(tree):
|
||||
"""Yield the expected report."""
|
||||
yield EXPECTED_REPORT + (type(plugin_func_gen), )
|
||||
yield EXPECTED_REPORT + (type(plugin_func_gen),)
|
||||
|
||||
|
||||
@plugin_func
|
||||
def plugin_func_list(tree):
|
||||
"""Return a list of expected reports."""
|
||||
return [EXPECTED_REPORT + (type(plugin_func_list), )]
|
||||
return [EXPECTED_REPORT + (type(plugin_func_list),)]
|
||||
|
||||
|
||||
@plugin_func
|
||||
|
|
@ -98,35 +98,37 @@ def mock_file_checker_with_plugin(plugin_target):
|
|||
Useful as a starting point for mocking reports/results.
|
||||
"""
|
||||
# Mock an entry point returning the plugin target
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
entry_point.name = plugin_target.name
|
||||
entry_point.load.return_value = plugin_target
|
||||
entry_point.value = 'mocked:value'
|
||||
entry_point.value = "mocked:value"
|
||||
|
||||
# Load the checker plugins using the entry point mock
|
||||
with mock.patch.object(
|
||||
importlib_metadata,
|
||||
'entry_points',
|
||||
return_value={'flake8.extension': [entry_point]},
|
||||
importlib_metadata,
|
||||
"entry_points",
|
||||
return_value={"flake8.extension": [entry_point]},
|
||||
):
|
||||
checks = manager.Checkers()
|
||||
|
||||
# Prevent it from reading lines from stdin or somewhere else
|
||||
with mock.patch('flake8.processor.FileProcessor.read_lines',
|
||||
return_value=['Line 1']):
|
||||
with mock.patch(
|
||||
"flake8.processor.FileProcessor.read_lines", return_value=["Line 1"]
|
||||
):
|
||||
file_checker = checker.FileChecker(
|
||||
'-',
|
||||
checks.to_dictionary(),
|
||||
mock.MagicMock()
|
||||
"-", checks.to_dictionary(), mock.MagicMock()
|
||||
)
|
||||
return file_checker
|
||||
|
||||
|
||||
@pytest.mark.parametrize('plugin_target', [
|
||||
PluginClass,
|
||||
plugin_func_gen,
|
||||
plugin_func_list,
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"plugin_target",
|
||||
[
|
||||
PluginClass,
|
||||
plugin_func_gen,
|
||||
plugin_func_list,
|
||||
],
|
||||
)
|
||||
def test_handle_file_plugins(plugin_target):
|
||||
"""Test the FileChecker class handling different file plugin types."""
|
||||
file_checker = mock_file_checker_with_plugin(plugin_target)
|
||||
|
|
@ -138,20 +140,25 @@ def test_handle_file_plugins(plugin_target):
|
|||
report = mock.Mock()
|
||||
file_checker.report = report
|
||||
file_checker.run_ast_checks()
|
||||
report.assert_called_once_with(error_code=None,
|
||||
line_number=EXPECTED_REPORT[0],
|
||||
column=EXPECTED_REPORT[1],
|
||||
text=EXPECTED_REPORT[2])
|
||||
report.assert_called_once_with(
|
||||
error_code=None,
|
||||
line_number=EXPECTED_REPORT[0],
|
||||
column=EXPECTED_REPORT[1],
|
||||
text=EXPECTED_REPORT[2],
|
||||
)
|
||||
|
||||
|
||||
@pytest.mark.parametrize('plugin_target,len_results', [
|
||||
(plugin_func_physical_ret, 1),
|
||||
(plugin_func_physical_none, 0),
|
||||
(plugin_func_physical_list_single, 1),
|
||||
(plugin_func_physical_list_multiple, 2),
|
||||
(plugin_func_physical_gen_single, 1),
|
||||
(plugin_func_physical_gen_multiple, 3),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"plugin_target,len_results",
|
||||
[
|
||||
(plugin_func_physical_ret, 1),
|
||||
(plugin_func_physical_none, 0),
|
||||
(plugin_func_physical_list_single, 1),
|
||||
(plugin_func_physical_list_multiple, 2),
|
||||
(plugin_func_physical_gen_single, 1),
|
||||
(plugin_func_physical_gen_multiple, 3),
|
||||
],
|
||||
)
|
||||
def test_line_check_results(plugin_target, len_results):
|
||||
"""Test the FileChecker class handling results from line checks."""
|
||||
file_checker = mock_file_checker_with_plugin(plugin_target)
|
||||
|
|
@ -167,54 +174,100 @@ def test_logical_line_offset_out_of_bounds():
|
|||
|
||||
@plugin_func
|
||||
def _logical_line_out_of_bounds(logical_line):
|
||||
yield 10000, 'L100 test'
|
||||
yield 10000, "L100 test"
|
||||
|
||||
file_checker = mock_file_checker_with_plugin(_logical_line_out_of_bounds)
|
||||
|
||||
logical_ret = (
|
||||
'',
|
||||
"",
|
||||
'print("xxxxxxxxxxx")',
|
||||
[(0, (1, 0)), (5, (1, 5)), (6, (1, 6)), (19, (1, 19)), (20, (1, 20))],
|
||||
)
|
||||
with mock.patch.object(
|
||||
FileProcessor, 'build_logical_line', return_value=logical_ret,
|
||||
FileProcessor,
|
||||
"build_logical_line",
|
||||
return_value=logical_ret,
|
||||
):
|
||||
file_checker.run_logical_checks()
|
||||
assert file_checker.results == [('L100', 0, 0, 'test', None)]
|
||||
assert file_checker.results == [("L100", 0, 0, "test", None)]
|
||||
|
||||
|
||||
PLACEHOLDER_CODE = 'some_line = "of" * code'
|
||||
|
||||
|
||||
@pytest.mark.parametrize('results, expected_order', [
|
||||
# No entries should be added
|
||||
([], []),
|
||||
# Results are correctly ordered
|
||||
([('A101', 1, 1, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 2, 1, 'placeholder error', PLACEHOLDER_CODE)], [0, 1]),
|
||||
# Reversed order of lines
|
||||
([('A101', 2, 1, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 1, 1, 'placeholder error', PLACEHOLDER_CODE)], [1, 0]),
|
||||
# Columns are not ordered correctly (when reports are ordered correctly)
|
||||
([('A101', 1, 2, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 1, 1, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 2, 1, 'placeholder error', PLACEHOLDER_CODE)], [1, 0, 2]),
|
||||
([('A101', 2, 1, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 1, 1, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 1, 2, 'placeholder error', PLACEHOLDER_CODE)], [1, 2, 0]),
|
||||
([('A101', 1, 2, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 2, 2, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 2, 1, 'placeholder error', PLACEHOLDER_CODE)], [0, 2, 1]),
|
||||
([('A101', 1, 3, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 2, 2, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 3, 1, 'placeholder error', PLACEHOLDER_CODE)], [0, 1, 2]),
|
||||
([('A101', 1, 1, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 1, 3, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 2, 2, 'placeholder error', PLACEHOLDER_CODE)], [0, 1, 2]),
|
||||
# Previously sort column and message (so reversed) (see bug 196)
|
||||
([('A101', 1, 1, 'placeholder error', PLACEHOLDER_CODE),
|
||||
('A101', 2, 1, 'charlie error', PLACEHOLDER_CODE)], [0, 1]),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"results, expected_order",
|
||||
[
|
||||
# No entries should be added
|
||||
([], []),
|
||||
# Results are correctly ordered
|
||||
(
|
||||
[
|
||||
("A101", 1, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 2, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[0, 1],
|
||||
),
|
||||
# Reversed order of lines
|
||||
(
|
||||
[
|
||||
("A101", 2, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 1, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[1, 0],
|
||||
),
|
||||
# Columns are not ordered correctly
|
||||
# (when reports are ordered correctly)
|
||||
(
|
||||
[
|
||||
("A101", 1, 2, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 1, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 2, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[1, 0, 2],
|
||||
),
|
||||
(
|
||||
[
|
||||
("A101", 2, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 1, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 1, 2, "placeholder error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[1, 2, 0],
|
||||
),
|
||||
(
|
||||
[
|
||||
("A101", 1, 2, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 2, 2, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 2, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[0, 2, 1],
|
||||
),
|
||||
(
|
||||
[
|
||||
("A101", 1, 3, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 2, 2, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 3, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[0, 1, 2],
|
||||
),
|
||||
(
|
||||
[
|
||||
("A101", 1, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 1, 3, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 2, 2, "placeholder error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[0, 1, 2],
|
||||
),
|
||||
# Previously sort column and message (so reversed) (see bug 196)
|
||||
(
|
||||
[
|
||||
("A101", 1, 1, "placeholder error", PLACEHOLDER_CODE),
|
||||
("A101", 2, 1, "charlie error", PLACEHOLDER_CODE),
|
||||
],
|
||||
[0, 1],
|
||||
),
|
||||
],
|
||||
)
|
||||
def test_report_order(results, expected_order):
|
||||
"""
|
||||
Test in which order the results will be reported.
|
||||
|
|
@ -222,6 +275,7 @@ def test_report_order(results, expected_order):
|
|||
It gets a list of reports from the file checkers and verifies that the
|
||||
result will be ordered independent from the original report.
|
||||
"""
|
||||
|
||||
def count_side_effect(name, sorted_results):
|
||||
"""Side effect for the result handler to tell all are reported."""
|
||||
return len(sorted_results)
|
||||
|
|
@ -230,11 +284,11 @@ def test_report_order(results, expected_order):
|
|||
# tuples to create the expected result lists from the indexes
|
||||
expected_results = [results[index] for index in expected_order]
|
||||
|
||||
file_checker = mock.Mock(spec=['results', 'display_name'])
|
||||
file_checker = mock.Mock(spec=["results", "display_name"])
|
||||
file_checker.results = results
|
||||
file_checker.display_name = 'placeholder'
|
||||
file_checker.display_name = "placeholder"
|
||||
|
||||
style_guide = mock.MagicMock(spec=['options', 'processing_file'])
|
||||
style_guide = mock.MagicMock(spec=["options", "processing_file"])
|
||||
|
||||
# Create a placeholder manager without arguments or plugins
|
||||
# Just add one custom file checker which just provides the results
|
||||
|
|
@ -244,9 +298,9 @@ def test_report_order(results, expected_order):
|
|||
# _handle_results is the first place which gets the sorted result
|
||||
# Should something non-private be mocked instead?
|
||||
handler = mock.Mock(side_effect=count_side_effect)
|
||||
with mock.patch.object(manager, '_handle_results', handler):
|
||||
with mock.patch.object(manager, "_handle_results", handler):
|
||||
assert manager.report() == (len(results), len(results))
|
||||
handler.assert_called_once_with('placeholder', expected_results)
|
||||
handler.assert_called_once_with("placeholder", expected_results)
|
||||
|
||||
|
||||
def test_acquire_when_multiprocessing_pool_can_initialize():
|
||||
|
|
|
|||
|
|
@ -17,7 +17,7 @@ def _call_main(argv, retv=0):
|
|||
|
||||
def test_diff_option(tmpdir, capsys):
|
||||
"""Ensure that `flake8 --diff` works."""
|
||||
t_py_contents = '''\
|
||||
t_py_contents = """\
|
||||
import os
|
||||
import sys # unused but not part of diff
|
||||
|
||||
|
|
@ -26,9 +26,9 @@ print('(to avoid trailing whitespace in test)')
|
|||
print(os.path.join('foo', 'bar'))
|
||||
|
||||
y # part of the diff and an error
|
||||
'''
|
||||
"""
|
||||
|
||||
diff = '''\
|
||||
diff = """\
|
||||
diff --git a/t.py b/t.py
|
||||
index d64ac39..7d943de 100644
|
||||
--- a/t.py
|
||||
|
|
@ -39,39 +39,39 @@ index d64ac39..7d943de 100644
|
|||
print(os.path.join('foo', 'bar'))
|
||||
+
|
||||
+y # part of the diff and an error
|
||||
'''
|
||||
"""
|
||||
|
||||
with mock.patch.object(utils, 'stdin_get_value', return_value=diff):
|
||||
with mock.patch.object(utils, "stdin_get_value", return_value=diff):
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write(t_py_contents)
|
||||
_call_main(['--diff'], retv=1)
|
||||
tmpdir.join("t.py").write(t_py_contents)
|
||||
_call_main(["--diff"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == "t.py:8:1: F821 undefined name 'y'\n"
|
||||
assert err == ''
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_form_feed_line_split(tmpdir, capsys):
|
||||
"""Test that form feed is treated the same for stdin."""
|
||||
src = 'x=1\n\f\ny=1\n'
|
||||
expected_out = '''\
|
||||
src = "x=1\n\f\ny=1\n"
|
||||
expected_out = """\
|
||||
t.py:1:2: E225 missing whitespace around operator
|
||||
t.py:3:2: E225 missing whitespace around operator
|
||||
'''
|
||||
"""
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write(src)
|
||||
tmpdir.join("t.py").write(src)
|
||||
|
||||
with mock.patch.object(utils, 'stdin_get_value', return_value=src):
|
||||
_call_main(['-', '--stdin-display-name=t.py'], retv=1)
|
||||
with mock.patch.object(utils, "stdin_get_value", return_value=src):
|
||||
_call_main(["-", "--stdin-display-name=t.py"], retv=1)
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected_out
|
||||
assert err == ''
|
||||
assert err == ""
|
||||
|
||||
_call_main(['t.py'], retv=1)
|
||||
_call_main(["t.py"], retv=1)
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected_out
|
||||
assert err == ''
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_e101_indent_char_does_not_reset(tmpdir, capsys):
|
||||
|
|
@ -89,82 +89,79 @@ if True:
|
|||
"""
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write(t_py_contents)
|
||||
_call_main(['t.py'])
|
||||
tmpdir.join("t.py").write(t_py_contents)
|
||||
_call_main(["t.py"])
|
||||
|
||||
|
||||
def test_statistics_option(tmpdir, capsys):
|
||||
"""Ensure that `flake8 --statistics` works."""
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write('import os\nimport sys\n')
|
||||
_call_main(['--statistics', 't.py'], retv=1)
|
||||
tmpdir.join("t.py").write("import os\nimport sys\n")
|
||||
_call_main(["--statistics", "t.py"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == '''\
|
||||
expected = """\
|
||||
t.py:1:1: F401 'os' imported but unused
|
||||
t.py:2:1: F401 'sys' imported but unused
|
||||
2 F401 'os' imported but unused
|
||||
'''
|
||||
assert err == ''
|
||||
"""
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_show_source_option(tmpdir, capsys):
|
||||
"""Ensure that --show-source and --no-show-source work."""
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('tox.ini').write('[flake8]\nshow_source = true\n')
|
||||
tmpdir.join('t.py').write('import os\n')
|
||||
_call_main(['t.py'], retv=1)
|
||||
tmpdir.join("tox.ini").write("[flake8]\nshow_source = true\n")
|
||||
tmpdir.join("t.py").write("import os\n")
|
||||
_call_main(["t.py"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == '''\
|
||||
expected = """\
|
||||
t.py:1:1: F401 'os' imported but unused
|
||||
import os
|
||||
^
|
||||
'''
|
||||
assert err == ''
|
||||
"""
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected
|
||||
assert err == ""
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
_call_main(['t.py', '--no-show-source'], retv=1)
|
||||
_call_main(["t.py", "--no-show-source"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == '''\
|
||||
expected = """\
|
||||
t.py:1:1: F401 'os' imported but unused
|
||||
'''
|
||||
assert err == ''
|
||||
"""
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_extend_exclude(tmpdir, capsys):
|
||||
"""Ensure that `flake8 --extend-exclude` works."""
|
||||
for d in ['project', 'vendor', 'legacy', '.git', '.tox', '.hg']:
|
||||
tmpdir.mkdir(d).join('t.py').write('import os\nimport sys\n')
|
||||
for d in ["project", "vendor", "legacy", ".git", ".tox", ".hg"]:
|
||||
tmpdir.mkdir(d).join("t.py").write("import os\nimport sys\n")
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
_call_main(['--extend-exclude=vendor,legacy/'], retv=1)
|
||||
_call_main(["--extend-exclude=vendor,legacy/"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
expected_out = '''\
|
||||
expected_out = """\
|
||||
./project/t.py:1:1: F401 'os' imported but unused
|
||||
./project/t.py:2:1: F401 'sys' imported but unused
|
||||
'''
|
||||
assert out == expected_out.replace('/', os.sep)
|
||||
assert err == ''
|
||||
"""
|
||||
assert out == expected_out.replace("/", os.sep)
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_malformed_per_file_ignores_error(tmpdir, capsys):
|
||||
"""Test the error message for malformed `per-file-ignores`."""
|
||||
setup_cfg = '''\
|
||||
setup_cfg = """\
|
||||
[flake8]
|
||||
per-file-ignores =
|
||||
incorrect/*
|
||||
values/*
|
||||
'''
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('setup.cfg').write(setup_cfg)
|
||||
_call_main(['.'], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == '''\
|
||||
"""
|
||||
expected = """\
|
||||
There was a critical error during execution of Flake8:
|
||||
Expected `per-file-ignores` to be a mapping from file exclude patterns to ignore codes.
|
||||
|
||||
|
|
@ -172,50 +169,59 @@ Configured `per-file-ignores` setting:
|
|||
|
||||
incorrect/*
|
||||
values/*
|
||||
''' # noqa: E501
|
||||
""" # noqa: E501
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join("setup.cfg").write(setup_cfg)
|
||||
_call_main(["."], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected
|
||||
|
||||
|
||||
def test_tokenization_error_but_not_syntax_error(tmpdir, capsys):
|
||||
"""Test that flake8 does not crash on tokenization errors."""
|
||||
with tmpdir.as_cwd():
|
||||
# this is a crash in the tokenizer, but not in the ast
|
||||
tmpdir.join('t.py').write("b'foo' \\\n")
|
||||
_call_main(['t.py'], retv=1)
|
||||
tmpdir.join("t.py").write("b'foo' \\\n")
|
||||
_call_main(["t.py"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == 't.py:1:1: E902 TokenError: EOF in multi-line statement\n'
|
||||
assert err == ''
|
||||
assert out == "t.py:1:1: E902 TokenError: EOF in multi-line statement\n"
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_tokenization_error_is_a_syntax_error(tmpdir, capsys):
|
||||
"""Test when tokenize raises a SyntaxError."""
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write('if True:\n pass\n pass\n')
|
||||
_call_main(['t.py'], retv=1)
|
||||
tmpdir.join("t.py").write("if True:\n pass\n pass\n")
|
||||
_call_main(["t.py"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == 't.py:1:1: E902 IndentationError: unindent does not match any outer indentation level\n' # noqa: E501
|
||||
assert err == ''
|
||||
expected = "t.py:1:1: E902 IndentationError: unindent does not match any outer indentation level\n" # noqa: E501
|
||||
assert out == expected
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_bug_report_successful(capsys):
|
||||
"""Test that --bug-report does not crash."""
|
||||
_call_main(['--bug-report'])
|
||||
_call_main(["--bug-report"])
|
||||
out, err = capsys.readouterr()
|
||||
assert json.loads(out)
|
||||
assert err == ''
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_specific_noqa_does_not_clobber_pycodestyle_noqa(tmpdir, capsys):
|
||||
"""See https://github.com/pycqa/flake8/issues/1104."""
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write("test = ('ABC' == None) # noqa: E501\n")
|
||||
_call_main(['t.py'], retv=1)
|
||||
tmpdir.join("t.py").write("test = ('ABC' == None) # noqa: E501\n")
|
||||
_call_main(["t.py"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == '''\
|
||||
expected = """\
|
||||
t.py:1:15: E711 comparison to None should be 'if cond is None:'
|
||||
'''
|
||||
"""
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected
|
||||
|
||||
|
||||
def test_specific_noqa_on_line_with_continuation(tmpdir, capsys):
|
||||
|
|
@ -230,60 +236,64 @@ x = """
|
|||
'''
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write(t_py_src)
|
||||
_call_main(['t.py'], retv=0)
|
||||
tmpdir.join("t.py").write(t_py_src)
|
||||
_call_main(["t.py"], retv=0)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == err == ''
|
||||
assert out == err == ""
|
||||
|
||||
|
||||
def test_physical_line_file_not_ending_in_newline(tmpdir, capsys):
|
||||
"""See https://github.com/PyCQA/pycodestyle/issues/960."""
|
||||
t_py_src = 'def f():\n\tpass'
|
||||
t_py_src = "def f():\n\tpass"
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write(t_py_src)
|
||||
_call_main(['t.py'], retv=1)
|
||||
tmpdir.join("t.py").write(t_py_src)
|
||||
_call_main(["t.py"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == '''\
|
||||
expected = """\
|
||||
t.py:2:1: W191 indentation contains tabs
|
||||
t.py:2:6: W292 no newline at end of file
|
||||
'''
|
||||
"""
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected
|
||||
|
||||
|
||||
def test_physical_line_file_not_ending_in_newline_trailing_ws(tmpdir, capsys):
|
||||
"""See https://github.com/PyCQA/pycodestyle/issues/960."""
|
||||
t_py_src = 'x = 1 '
|
||||
t_py_src = "x = 1 "
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('t.py').write(t_py_src)
|
||||
_call_main(['t.py'], retv=1)
|
||||
tmpdir.join("t.py").write(t_py_src)
|
||||
_call_main(["t.py"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == '''\
|
||||
expected = """\
|
||||
t.py:1:6: W291 trailing whitespace
|
||||
t.py:1:9: W292 no newline at end of file
|
||||
'''
|
||||
"""
|
||||
out, err = capsys.readouterr()
|
||||
assert out == expected
|
||||
|
||||
|
||||
def test_obtaining_args_from_sys_argv_when_not_explicity_provided(capsys):
|
||||
"""Test that arguments are obtained from 'sys.argv'."""
|
||||
with mock.patch('sys.argv', ['flake8', '--help']):
|
||||
with mock.patch("sys.argv", ["flake8", "--help"]):
|
||||
_call_main(None)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out.startswith('usage: flake8 [options] file file ...\n')
|
||||
assert err == ''
|
||||
assert out.startswith("usage: flake8 [options] file file ...\n")
|
||||
assert err == ""
|
||||
|
||||
|
||||
def test_cli_config_option_respected(tmp_path):
|
||||
"""Test --config is used."""
|
||||
config = tmp_path / "flake8.ini"
|
||||
config.write_text("""\
|
||||
config.write_text(
|
||||
"""\
|
||||
[flake8]
|
||||
ignore = F401
|
||||
""")
|
||||
"""
|
||||
)
|
||||
|
||||
py_file = tmp_path / "t.py"
|
||||
py_file.write_text("import os\n")
|
||||
|
|
@ -294,10 +304,12 @@ ignore = F401
|
|||
def test_cli_isolated_overrides_config_option(tmp_path):
|
||||
"""Test --isolated overrides --config."""
|
||||
config = tmp_path / "flake8.ini"
|
||||
config.write_text("""\
|
||||
config.write_text(
|
||||
"""\
|
||||
[flake8]
|
||||
ignore = F401
|
||||
""")
|
||||
"""
|
||||
)
|
||||
|
||||
py_file = tmp_path / "t.py"
|
||||
py_file.write_text("import os\n")
|
||||
|
|
@ -316,13 +328,13 @@ def test_file_not_found(tmpdir, capsys):
|
|||
|
||||
def test_output_file(tmpdir, capsys):
|
||||
"""Ensure that --output-file is honored."""
|
||||
tmpdir.join('t.py').write('import os\n')
|
||||
tmpdir.join("t.py").write("import os\n")
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
_call_main(['t.py', '--output-file=f'], retv=1)
|
||||
_call_main(["t.py", "--output-file=f"], retv=1)
|
||||
|
||||
out, err = capsys.readouterr()
|
||||
assert out == err == ""
|
||||
|
||||
expected = "t.py:1:1: F401 'os' imported but unused\n"
|
||||
assert tmpdir.join('f').read() == expected
|
||||
assert tmpdir.join("f").read() == expected
|
||||
|
|
|
|||
|
|
@ -1,15 +1,15 @@
|
|||
"""Integration tests for plugin loading."""
|
||||
from flake8.main import application
|
||||
|
||||
LOCAL_PLUGIN_CONFIG = 'tests/fixtures/config_files/local-plugin.ini'
|
||||
LOCAL_PLUGIN_PATH_CONFIG = 'tests/fixtures/config_files/local-plugin-path.ini'
|
||||
LOCAL_PLUGIN_CONFIG = "tests/fixtures/config_files/local-plugin.ini"
|
||||
LOCAL_PLUGIN_PATH_CONFIG = "tests/fixtures/config_files/local-plugin-path.ini"
|
||||
|
||||
|
||||
class ExtensionTestPlugin:
|
||||
"""Extension test plugin."""
|
||||
|
||||
name = 'ExtensionTestPlugin'
|
||||
version = '1.0.0'
|
||||
name = "ExtensionTestPlugin"
|
||||
version = "1.0.0"
|
||||
|
||||
def __init__(self, tree):
|
||||
"""Construct an instance of test plugin."""
|
||||
|
|
@ -20,14 +20,14 @@ class ExtensionTestPlugin:
|
|||
@classmethod
|
||||
def add_options(cls, parser):
|
||||
"""Register options."""
|
||||
parser.add_option('--anopt')
|
||||
parser.add_option("--anopt")
|
||||
|
||||
|
||||
class ReportTestPlugin:
|
||||
"""Report test plugin."""
|
||||
|
||||
name = 'ReportTestPlugin'
|
||||
version = '1.0.0'
|
||||
name = "ReportTestPlugin"
|
||||
version = "1.0.0"
|
||||
|
||||
def __init__(self, tree):
|
||||
"""Construct an instance of test plugin."""
|
||||
|
|
@ -39,28 +39,29 @@ class ReportTestPlugin:
|
|||
def test_enable_local_plugin_from_config():
|
||||
"""App can load a local plugin from config file."""
|
||||
app = application.Application()
|
||||
app.initialize(['flake8', '--config', LOCAL_PLUGIN_CONFIG])
|
||||
app.initialize(["flake8", "--config", LOCAL_PLUGIN_CONFIG])
|
||||
|
||||
assert app.check_plugins is not None
|
||||
assert app.check_plugins['XE'].plugin is ExtensionTestPlugin
|
||||
assert app.check_plugins["XE"].plugin is ExtensionTestPlugin
|
||||
assert app.formatting_plugins is not None
|
||||
assert app.formatting_plugins['XR'].plugin is ReportTestPlugin
|
||||
assert app.formatting_plugins["XR"].plugin is ReportTestPlugin
|
||||
|
||||
|
||||
def test_local_plugin_can_add_option():
|
||||
"""A local plugin can add a CLI option."""
|
||||
app = application.Application()
|
||||
app.initialize(
|
||||
['flake8', '--config', LOCAL_PLUGIN_CONFIG, '--anopt', 'foo'])
|
||||
["flake8", "--config", LOCAL_PLUGIN_CONFIG, "--anopt", "foo"]
|
||||
)
|
||||
|
||||
assert app.options is not None
|
||||
assert app.options.anopt == 'foo'
|
||||
assert app.options.anopt == "foo"
|
||||
|
||||
|
||||
def test_enable_local_plugin_at_non_installed_path():
|
||||
"""Can add a paths option in local-plugins config section for finding."""
|
||||
app = application.Application()
|
||||
app.initialize(['flake8', '--config', LOCAL_PLUGIN_PATH_CONFIG])
|
||||
app.initialize(["flake8", "--config", LOCAL_PLUGIN_PATH_CONFIG])
|
||||
|
||||
assert app.check_plugins is not None
|
||||
assert app.check_plugins['XE'].plugin.name == 'ExtensionTestPlugin2'
|
||||
assert app.check_plugins["XE"].plugin.name == "ExtensionTestPlugin2"
|
||||
|
|
|
|||
|
|
@ -6,13 +6,13 @@ import pytest
|
|||
|
||||
def options_from(**kwargs):
|
||||
"""Generate a Values instances with our kwargs."""
|
||||
kwargs.setdefault('hang_closing', True)
|
||||
kwargs.setdefault('max_line_length', 79)
|
||||
kwargs.setdefault('max_doc_length', None)
|
||||
kwargs.setdefault('indent_size', 4)
|
||||
kwargs.setdefault('verbose', False)
|
||||
kwargs.setdefault('stdin_display_name', 'stdin')
|
||||
kwargs.setdefault('disable_noqa', False)
|
||||
kwargs.setdefault("hang_closing", True)
|
||||
kwargs.setdefault("max_line_length", 79)
|
||||
kwargs.setdefault("max_doc_length", None)
|
||||
kwargs.setdefault("indent_size", 4)
|
||||
kwargs.setdefault("verbose", False)
|
||||
kwargs.setdefault("stdin_display_name", "stdin")
|
||||
kwargs.setdefault("disable_noqa", False)
|
||||
return argparse.Namespace(**kwargs)
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -10,10 +10,10 @@ from flake8.main import application as app
|
|||
|
||||
def options(**kwargs):
|
||||
"""Generate argparse.Namespace for our Application."""
|
||||
kwargs.setdefault('verbose', 0)
|
||||
kwargs.setdefault('output_file', None)
|
||||
kwargs.setdefault('count', False)
|
||||
kwargs.setdefault('exit_zero', False)
|
||||
kwargs.setdefault("verbose", 0)
|
||||
kwargs.setdefault("output_file", None)
|
||||
kwargs.setdefault("count", False)
|
||||
kwargs.setdefault("exit_zero", False)
|
||||
return argparse.Namespace(**kwargs)
|
||||
|
||||
|
||||
|
|
@ -24,19 +24,20 @@ def application():
|
|||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'result_count, catastrophic, exit_zero, value', [
|
||||
"result_count, catastrophic, exit_zero, value",
|
||||
[
|
||||
(0, False, False, False),
|
||||
(0, True, False, True),
|
||||
(2, False, False, True),
|
||||
(2, True, False, True),
|
||||
|
||||
(0, True, True, True),
|
||||
(2, False, True, False),
|
||||
(2, True, True, True),
|
||||
]
|
||||
],
|
||||
)
|
||||
def test_exit_does_raise(result_count, catastrophic, exit_zero, value,
|
||||
application):
|
||||
def test_exit_does_raise(
|
||||
result_count, catastrophic, exit_zero, value, application
|
||||
):
|
||||
"""Verify Application.exit doesn't raise SystemExit."""
|
||||
application.result_count = result_count
|
||||
application.catastrophic_failure = catastrophic
|
||||
|
|
@ -53,10 +54,10 @@ def test_warns_on_unknown_formatter_plugin_name(application):
|
|||
default = mock.Mock()
|
||||
execute = default.execute
|
||||
application.formatting_plugins = {
|
||||
'default': default,
|
||||
"default": default,
|
||||
}
|
||||
with mock.patch.object(app.LOG, 'warning') as warning:
|
||||
assert execute is application.formatter_for('fake-plugin-name')
|
||||
with mock.patch.object(app.LOG, "warning") as warning:
|
||||
assert execute is application.formatter_for("fake-plugin-name")
|
||||
|
||||
assert warning.called is True
|
||||
assert warning.call_count == 1
|
||||
|
|
@ -67,12 +68,12 @@ def test_returns_specified_plugin(application):
|
|||
desired = mock.Mock()
|
||||
execute = desired.execute
|
||||
application.formatting_plugins = {
|
||||
'default': mock.Mock(),
|
||||
'desired': desired,
|
||||
"default": mock.Mock(),
|
||||
"desired": desired,
|
||||
}
|
||||
|
||||
with mock.patch.object(app.LOG, 'warning') as warning:
|
||||
assert execute is application.formatter_for('desired')
|
||||
with mock.patch.object(app.LOG, "warning") as warning:
|
||||
assert execute is application.formatter_for("desired")
|
||||
|
||||
assert warning.called is False
|
||||
|
||||
|
|
@ -80,10 +81,11 @@ def test_returns_specified_plugin(application):
|
|||
def test_prelim_opts_args(application):
|
||||
"""Verify we get sensible prelim opts and args."""
|
||||
opts, args = application.parse_preliminary_options(
|
||||
['--foo', '--verbose', 'src', 'setup.py', '--statistics', '--version'])
|
||||
["--foo", "--verbose", "src", "setup.py", "--statistics", "--version"]
|
||||
)
|
||||
|
||||
assert opts.verbose
|
||||
assert args == ['--foo', 'src', 'setup.py', '--statistics', '--version']
|
||||
assert args == ["--foo", "src", "setup.py", "--statistics", "--version"]
|
||||
|
||||
|
||||
def test_prelim_opts_ignore_help(application):
|
||||
|
|
@ -91,16 +93,16 @@ def test_prelim_opts_ignore_help(application):
|
|||
# GIVEN
|
||||
|
||||
# WHEN
|
||||
_, args = application.parse_preliminary_options(['--help', '-h'])
|
||||
_, args = application.parse_preliminary_options(["--help", "-h"])
|
||||
|
||||
# THEN
|
||||
assert args == ['--help', '-h']
|
||||
assert args == ["--help", "-h"]
|
||||
|
||||
|
||||
def test_prelim_opts_handles_empty(application):
|
||||
"""Verify empty argv lists are handled correctly."""
|
||||
irrelevant_args = ['myexe', '/path/to/foo']
|
||||
with mock.patch.object(sys, 'argv', irrelevant_args):
|
||||
irrelevant_args = ["myexe", "/path/to/foo"]
|
||||
with mock.patch.object(sys, "argv", irrelevant_args):
|
||||
opts, args = application.parse_preliminary_options([])
|
||||
|
||||
assert args == []
|
||||
|
|
|
|||
|
|
@ -10,23 +10,23 @@ from flake8.formatting import base
|
|||
|
||||
def options(**kwargs):
|
||||
"""Create an argparse.Namespace instance."""
|
||||
kwargs.setdefault('output_file', None)
|
||||
kwargs.setdefault('tee', False)
|
||||
kwargs.setdefault("output_file", None)
|
||||
kwargs.setdefault("tee", False)
|
||||
return argparse.Namespace(**kwargs)
|
||||
|
||||
|
||||
@pytest.mark.parametrize('filename', [None, 'out.txt'])
|
||||
@pytest.mark.parametrize("filename", [None, "out.txt"])
|
||||
def test_start(filename):
|
||||
"""Verify we open a new file in the start method."""
|
||||
mock_open = mock.mock_open()
|
||||
formatter = base.BaseFormatter(options(output_file=filename))
|
||||
with mock.patch('flake8.formatting.base.open', mock_open):
|
||||
with mock.patch("flake8.formatting.base.open", mock_open):
|
||||
formatter.start()
|
||||
|
||||
if filename is None:
|
||||
assert mock_open.called is False
|
||||
else:
|
||||
mock_open.assert_called_once_with(filename, 'a')
|
||||
mock_open.assert_called_once_with(filename, "a")
|
||||
|
||||
|
||||
def test_stop():
|
||||
|
|
@ -45,68 +45,79 @@ def test_format_needs_to_be_implemented():
|
|||
formatter = base.BaseFormatter(options())
|
||||
with pytest.raises(NotImplementedError):
|
||||
formatter.format(
|
||||
style_guide.Violation('A000', 'file.py', 1, 1, 'error text', None)
|
||||
style_guide.Violation("A000", "file.py", 1, 1, "error text", None)
|
||||
)
|
||||
|
||||
|
||||
def test_show_source_returns_nothing_when_not_showing_source():
|
||||
"""Ensure we return nothing when users want nothing."""
|
||||
formatter = base.BaseFormatter(options(show_source=False))
|
||||
assert formatter.show_source(
|
||||
style_guide.Violation('A000', 'file.py', 1, 1, 'error text', 'line')
|
||||
) == ''
|
||||
assert (
|
||||
formatter.show_source(
|
||||
style_guide.Violation(
|
||||
"A000", "file.py", 1, 1, "error text", "line"
|
||||
)
|
||||
)
|
||||
== ""
|
||||
)
|
||||
|
||||
|
||||
def test_show_source_returns_nothing_when_there_is_source():
|
||||
"""Ensure we return nothing when there is no line."""
|
||||
formatter = base.BaseFormatter(options(show_source=True))
|
||||
assert formatter.show_source(
|
||||
style_guide.Violation('A000', 'file.py', 1, 1, 'error text', None)
|
||||
) == ''
|
||||
assert (
|
||||
formatter.show_source(
|
||||
style_guide.Violation("A000", "file.py", 1, 1, "error text", None)
|
||||
)
|
||||
== ""
|
||||
)
|
||||
|
||||
|
||||
@pytest.mark.parametrize(('line1', 'line2', 'column'), [
|
||||
(
|
||||
'x=1\n',
|
||||
' ^',
|
||||
2,
|
||||
),
|
||||
(
|
||||
' x=(1\n +2)\n',
|
||||
' ^',
|
||||
5,
|
||||
),
|
||||
(
|
||||
'\tx\t=\ty\n',
|
||||
'\t \t \t^',
|
||||
6,
|
||||
),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
("line1", "line2", "column"),
|
||||
[
|
||||
(
|
||||
"x=1\n",
|
||||
" ^",
|
||||
2,
|
||||
),
|
||||
(
|
||||
" x=(1\n +2)\n",
|
||||
" ^",
|
||||
5,
|
||||
),
|
||||
(
|
||||
"\tx\t=\ty\n",
|
||||
"\t \t \t^",
|
||||
6,
|
||||
),
|
||||
],
|
||||
)
|
||||
def test_show_source_updates_physical_line_appropriately(line1, line2, column):
|
||||
"""Ensure the error column is appropriately indicated."""
|
||||
formatter = base.BaseFormatter(options(show_source=True))
|
||||
error = style_guide.Violation('A000', 'file.py', 1, column, 'error', line1)
|
||||
error = style_guide.Violation("A000", "file.py", 1, column, "error", line1)
|
||||
output = formatter.show_source(error)
|
||||
assert output == line1 + line2
|
||||
|
||||
|
||||
@pytest.mark.parametrize('tee', [False, True])
|
||||
@pytest.mark.parametrize("tee", [False, True])
|
||||
def test_write_uses_an_output_file(tee):
|
||||
"""Verify that we use the output file when it's present."""
|
||||
line = 'Something to write'
|
||||
source = 'source'
|
||||
line = "Something to write"
|
||||
source = "source"
|
||||
filemock = mock.Mock()
|
||||
|
||||
formatter = base.BaseFormatter(options(tee=tee))
|
||||
formatter.output_fd = filemock
|
||||
|
||||
with mock.patch('flake8.formatting.base.print') as print_func:
|
||||
with mock.patch("flake8.formatting.base.print") as print_func:
|
||||
formatter.write(line, source)
|
||||
if tee:
|
||||
assert print_func.called
|
||||
assert print_func.mock_calls == [
|
||||
mock.call(line, end='\n'),
|
||||
mock.call(source, end='\n'),
|
||||
mock.call(line, end="\n"),
|
||||
mock.call(source, end="\n"),
|
||||
]
|
||||
else:
|
||||
assert not print_func.called
|
||||
|
|
@ -119,11 +130,11 @@ def test_write_uses_an_output_file(tee):
|
|||
]
|
||||
|
||||
|
||||
@mock.patch('flake8.formatting.base.print')
|
||||
@mock.patch("flake8.formatting.base.print")
|
||||
def test_write_uses_print(print_function):
|
||||
"""Verify that we use the print function without an output file."""
|
||||
line = 'Something to write'
|
||||
source = 'source'
|
||||
line = "Something to write"
|
||||
source = "source"
|
||||
|
||||
formatter = base.BaseFormatter(options())
|
||||
formatter.write(line, source)
|
||||
|
|
@ -131,8 +142,8 @@ def test_write_uses_print(print_function):
|
|||
assert print_function.called is True
|
||||
assert print_function.call_count == 2
|
||||
assert print_function.mock_calls == [
|
||||
mock.call(line, end='\n'),
|
||||
mock.call(source, end='\n'),
|
||||
mock.call(line, end="\n"),
|
||||
mock.call(source, end="\n"),
|
||||
]
|
||||
|
||||
|
||||
|
|
@ -163,14 +174,14 @@ def test_handle_formats_the_error():
|
|||
formatter = FormatFormatter(options(show_source=False))
|
||||
filemock = formatter.output_fd = mock.Mock()
|
||||
error = style_guide.Violation(
|
||||
code='A001',
|
||||
filename='example.py',
|
||||
code="A001",
|
||||
filename="example.py",
|
||||
line_number=1,
|
||||
column_number=1,
|
||||
text='Fake error',
|
||||
physical_line='a = 1',
|
||||
text="Fake error",
|
||||
physical_line="a = 1",
|
||||
)
|
||||
|
||||
formatter.handle(error)
|
||||
|
||||
filemock.write.assert_called_once_with(repr(error) + '\n')
|
||||
filemock.write.assert_called_once_with(repr(error) + "\n")
|
||||
|
|
|
|||
|
|
@ -10,10 +10,12 @@ from flake8.main.options import JobsArgument
|
|||
|
||||
def style_guide_mock():
|
||||
"""Create a mock StyleGuide object."""
|
||||
return mock.MagicMock(**{
|
||||
'options.diff': False,
|
||||
'options.jobs': JobsArgument("4"),
|
||||
})
|
||||
return mock.MagicMock(
|
||||
**{
|
||||
"options.diff": False,
|
||||
"options.jobs": JobsArgument("4"),
|
||||
}
|
||||
)
|
||||
|
||||
|
||||
def _parallel_checker_manager():
|
||||
|
|
@ -27,21 +29,21 @@ def _parallel_checker_manager():
|
|||
|
||||
def test_oserrors_cause_serial_fall_back():
|
||||
"""Verify that OSErrors will cause the Manager to fallback to serial."""
|
||||
err = OSError(errno.ENOSPC, 'Ominous message about spaceeeeee')
|
||||
with mock.patch('_multiprocessing.SemLock', side_effect=err):
|
||||
err = OSError(errno.ENOSPC, "Ominous message about spaceeeeee")
|
||||
with mock.patch("_multiprocessing.SemLock", side_effect=err):
|
||||
manager = _parallel_checker_manager()
|
||||
with mock.patch.object(manager, 'run_serial') as serial:
|
||||
with mock.patch.object(manager, "run_serial") as serial:
|
||||
manager.run()
|
||||
assert serial.call_count == 1
|
||||
|
||||
|
||||
@mock.patch('flake8.checker._multiprocessing_is_fork', return_value=True)
|
||||
@mock.patch("flake8.checker._multiprocessing_is_fork", return_value=True)
|
||||
def test_oserrors_are_reraised(is_windows):
|
||||
"""Verify that unexpected OSErrors will cause the Manager to reraise."""
|
||||
err = OSError(errno.EAGAIN, 'Ominous message')
|
||||
with mock.patch('_multiprocessing.SemLock', side_effect=err):
|
||||
err = OSError(errno.EAGAIN, "Ominous message")
|
||||
with mock.patch("_multiprocessing.SemLock", side_effect=err):
|
||||
manager = _parallel_checker_manager()
|
||||
with mock.patch.object(manager, 'run_serial') as serial:
|
||||
with mock.patch.object(manager, "run_serial") as serial:
|
||||
with pytest.raises(OSError):
|
||||
manager.run()
|
||||
assert serial.call_count == 0
|
||||
|
|
@ -50,7 +52,7 @@ def test_oserrors_are_reraised(is_windows):
|
|||
def test_multiprocessing_is_disabled():
|
||||
"""Verify not being able to import multiprocessing forces jobs to 0."""
|
||||
style_guide = style_guide_mock()
|
||||
with mock.patch('flake8.checker.multiprocessing', None):
|
||||
with mock.patch("flake8.checker.multiprocessing", None):
|
||||
manager = checker.Manager(style_guide, [], [])
|
||||
assert manager.jobs == 0
|
||||
|
||||
|
|
@ -58,20 +60,20 @@ def test_multiprocessing_is_disabled():
|
|||
def test_make_checkers():
|
||||
"""Verify that we create a list of FileChecker instances."""
|
||||
style_guide = style_guide_mock()
|
||||
files = ['file1', 'file2']
|
||||
files = ["file1", "file2"]
|
||||
checkplugins = mock.Mock()
|
||||
checkplugins.to_dictionary.return_value = {
|
||||
'ast_plugins': [],
|
||||
'logical_line_plugins': [],
|
||||
'physical_line_plugins': [],
|
||||
"ast_plugins": [],
|
||||
"logical_line_plugins": [],
|
||||
"physical_line_plugins": [],
|
||||
}
|
||||
with mock.patch('flake8.checker.multiprocessing', None):
|
||||
with mock.patch("flake8.checker.multiprocessing", None):
|
||||
manager = checker.Manager(style_guide, files, checkplugins)
|
||||
|
||||
with mock.patch('flake8.utils.filenames_from') as filenames_from:
|
||||
filenames_from.side_effect = [['file1'], ['file2']]
|
||||
with mock.patch('flake8.utils.fnmatch', return_value=True):
|
||||
with mock.patch('flake8.processor.FileProcessor'):
|
||||
with mock.patch("flake8.utils.filenames_from") as filenames_from:
|
||||
filenames_from.side_effect = [["file1"], ["file2"]]
|
||||
with mock.patch("flake8.utils.fnmatch", return_value=True):
|
||||
with mock.patch("flake8.processor.FileProcessor"):
|
||||
manager.make_checkers()
|
||||
|
||||
assert manager._all_checkers
|
||||
|
|
|
|||
|
|
@ -7,73 +7,93 @@ import pytest
|
|||
|
||||
from flake8.options import config
|
||||
|
||||
CLI_SPECIFIED_FILEPATH = 'tests/fixtures/config_files/cli-specified.ini'
|
||||
BROKEN_CONFIG_PATH = 'tests/fixtures/config_files/broken.ini'
|
||||
CLI_SPECIFIED_FILEPATH = "tests/fixtures/config_files/cli-specified.ini"
|
||||
BROKEN_CONFIG_PATH = "tests/fixtures/config_files/broken.ini"
|
||||
|
||||
|
||||
def test_cli_config():
|
||||
"""Verify opening and reading the file specified via the cli."""
|
||||
cli_filepath = CLI_SPECIFIED_FILEPATH
|
||||
finder = config.ConfigFileFinder('flake8')
|
||||
finder = config.ConfigFileFinder("flake8")
|
||||
|
||||
parsed_config = finder.cli_config(cli_filepath)
|
||||
assert parsed_config.has_section('flake8')
|
||||
assert parsed_config.has_section("flake8")
|
||||
|
||||
|
||||
@pytest.mark.parametrize('cwd,expected', [
|
||||
# Root directory of project
|
||||
(os.path.abspath('.'),
|
||||
[os.path.abspath('setup.cfg'),
|
||||
os.path.abspath('tox.ini')]),
|
||||
# Subdirectory of project directory
|
||||
(os.path.abspath('src'),
|
||||
[os.path.abspath('setup.cfg'),
|
||||
os.path.abspath('tox.ini')]),
|
||||
# Outside of project directory
|
||||
(os.path.abspath('/'),
|
||||
[]),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"cwd,expected",
|
||||
[
|
||||
# Root directory of project
|
||||
(
|
||||
os.path.abspath("."),
|
||||
[os.path.abspath("setup.cfg"), os.path.abspath("tox.ini")],
|
||||
),
|
||||
# Subdirectory of project directory
|
||||
(
|
||||
os.path.abspath("src"),
|
||||
[os.path.abspath("setup.cfg"), os.path.abspath("tox.ini")],
|
||||
),
|
||||
# Outside of project directory
|
||||
(os.path.abspath("/"), []),
|
||||
],
|
||||
)
|
||||
def test_generate_possible_local_files(cwd, expected):
|
||||
"""Verify generation of all possible config paths."""
|
||||
finder = config.ConfigFileFinder('flake8')
|
||||
finder = config.ConfigFileFinder("flake8")
|
||||
|
||||
with mock.patch.object(os, 'getcwd', return_value=cwd):
|
||||
with mock.patch.object(os, "getcwd", return_value=cwd):
|
||||
config_files = list(finder.generate_possible_local_files())
|
||||
|
||||
assert config_files == expected
|
||||
|
||||
|
||||
@pytest.mark.parametrize('extra_config_files,expected', [
|
||||
# Extra config files specified
|
||||
([CLI_SPECIFIED_FILEPATH],
|
||||
[os.path.abspath('setup.cfg'),
|
||||
os.path.abspath('tox.ini'),
|
||||
os.path.abspath(CLI_SPECIFIED_FILEPATH)]),
|
||||
# Missing extra config files specified
|
||||
([CLI_SPECIFIED_FILEPATH,
|
||||
'tests/fixtures/config_files/missing.ini'],
|
||||
[os.path.abspath('setup.cfg'),
|
||||
os.path.abspath('tox.ini'),
|
||||
os.path.abspath(CLI_SPECIFIED_FILEPATH)]),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"extra_config_files,expected",
|
||||
[
|
||||
# Extra config files specified
|
||||
(
|
||||
[CLI_SPECIFIED_FILEPATH],
|
||||
[
|
||||
os.path.abspath("setup.cfg"),
|
||||
os.path.abspath("tox.ini"),
|
||||
os.path.abspath(CLI_SPECIFIED_FILEPATH),
|
||||
],
|
||||
),
|
||||
# Missing extra config files specified
|
||||
(
|
||||
[
|
||||
CLI_SPECIFIED_FILEPATH,
|
||||
"tests/fixtures/config_files/missing.ini",
|
||||
],
|
||||
[
|
||||
os.path.abspath("setup.cfg"),
|
||||
os.path.abspath("tox.ini"),
|
||||
os.path.abspath(CLI_SPECIFIED_FILEPATH),
|
||||
],
|
||||
),
|
||||
],
|
||||
)
|
||||
def test_local_config_files(extra_config_files, expected):
|
||||
"""Verify discovery of local config files."""
|
||||
finder = config.ConfigFileFinder('flake8', extra_config_files)
|
||||
finder = config.ConfigFileFinder("flake8", extra_config_files)
|
||||
|
||||
assert list(finder.local_config_files()) == expected
|
||||
|
||||
|
||||
def test_local_configs():
|
||||
"""Verify we return a ConfigParser."""
|
||||
finder = config.ConfigFileFinder('flake8')
|
||||
finder = config.ConfigFileFinder("flake8")
|
||||
|
||||
assert isinstance(finder.local_configs(), configparser.RawConfigParser)
|
||||
|
||||
|
||||
@pytest.mark.parametrize('files', [
|
||||
[BROKEN_CONFIG_PATH],
|
||||
[CLI_SPECIFIED_FILEPATH, BROKEN_CONFIG_PATH],
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"files",
|
||||
[
|
||||
[BROKEN_CONFIG_PATH],
|
||||
[CLI_SPECIFIED_FILEPATH, BROKEN_CONFIG_PATH],
|
||||
],
|
||||
)
|
||||
def test_read_config_catches_broken_config_files(files):
|
||||
"""Verify that we do not allow the exception to bubble up."""
|
||||
_, parsed = config.ConfigFileFinder._read_config(*files)
|
||||
|
|
@ -82,40 +102,42 @@ def test_read_config_catches_broken_config_files(files):
|
|||
|
||||
def test_read_config_catches_decoding_errors(tmpdir):
|
||||
"""Verify that we do not allow the exception to bubble up."""
|
||||
setup_cfg = tmpdir.join('setup.cfg')
|
||||
setup_cfg = tmpdir.join("setup.cfg")
|
||||
# pick bytes that are unlikely to decode
|
||||
setup_cfg.write_binary(b'[x]\ny = \x81\x8d\x90\x9d')
|
||||
setup_cfg.write_binary(b"[x]\ny = \x81\x8d\x90\x9d")
|
||||
_, parsed = config.ConfigFileFinder._read_config(setup_cfg.strpath)
|
||||
assert parsed == []
|
||||
|
||||
|
||||
def test_config_file_default_value():
|
||||
"""Verify the default 'config_file' attribute value."""
|
||||
finder = config.ConfigFileFinder('flake8')
|
||||
finder = config.ConfigFileFinder("flake8")
|
||||
assert finder.config_file is None
|
||||
|
||||
|
||||
def test_setting_config_file_value():
|
||||
"""Verify the 'config_file' attribute matches constructed value."""
|
||||
config_file_value = 'flake8.ini'
|
||||
finder = config.ConfigFileFinder('flake8', config_file=config_file_value)
|
||||
config_file_value = "flake8.ini"
|
||||
finder = config.ConfigFileFinder("flake8", config_file=config_file_value)
|
||||
assert finder.config_file == config_file_value
|
||||
|
||||
|
||||
def test_ignore_config_files_default_value():
|
||||
"""Verify the default 'ignore_config_files' attribute value."""
|
||||
finder = config.ConfigFileFinder('flake8')
|
||||
finder = config.ConfigFileFinder("flake8")
|
||||
assert finder.ignore_config_files is False
|
||||
|
||||
|
||||
@pytest.mark.parametrize('ignore_config_files_arg', [
|
||||
False,
|
||||
True,
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"ignore_config_files_arg",
|
||||
[
|
||||
False,
|
||||
True,
|
||||
],
|
||||
)
|
||||
def test_setting_ignore_config_files_value(ignore_config_files_arg):
|
||||
"""Verify the 'ignore_config_files' attribute matches constructed value."""
|
||||
finder = config.ConfigFileFinder(
|
||||
'flake8',
|
||||
ignore_config_files=ignore_config_files_arg
|
||||
"flake8", ignore_config_files=ignore_config_files_arg
|
||||
)
|
||||
assert finder.ignore_config_files is ignore_config_files_arg
|
||||
|
|
|
|||
|
|
@ -12,51 +12,82 @@ def test_dependencies():
|
|||
assert [] == debug.dependencies()
|
||||
|
||||
|
||||
@pytest.mark.parametrize('plugins, expected', [
|
||||
([], []),
|
||||
([manager.PluginVersion('pycodestyle', '2.0.0', False)],
|
||||
[{'plugin': 'pycodestyle', 'version': '2.0.0', 'is_local': False}]),
|
||||
([manager.PluginVersion('pycodestyle', '2.0.0', False),
|
||||
manager.PluginVersion('mccabe', '0.5.9', False)],
|
||||
[{'plugin': 'mccabe', 'version': '0.5.9', 'is_local': False},
|
||||
{'plugin': 'pycodestyle', 'version': '2.0.0', 'is_local': False}]),
|
||||
([manager.PluginVersion('pycodestyle', '2.0.0', False),
|
||||
manager.PluginVersion('my-local', '0.0.1', True),
|
||||
manager.PluginVersion('mccabe', '0.5.9', False)],
|
||||
[{'plugin': 'mccabe', 'version': '0.5.9', 'is_local': False},
|
||||
{'plugin': 'my-local', 'version': '0.0.1', 'is_local': True},
|
||||
{'plugin': 'pycodestyle', 'version': '2.0.0', 'is_local': False}]),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"plugins, expected",
|
||||
[
|
||||
([], []),
|
||||
(
|
||||
[manager.PluginVersion("pycodestyle", "2.0.0", False)],
|
||||
[
|
||||
{
|
||||
"plugin": "pycodestyle",
|
||||
"version": "2.0.0",
|
||||
"is_local": False,
|
||||
}
|
||||
],
|
||||
),
|
||||
(
|
||||
[
|
||||
manager.PluginVersion("pycodestyle", "2.0.0", False),
|
||||
manager.PluginVersion("mccabe", "0.5.9", False),
|
||||
],
|
||||
[
|
||||
{"plugin": "mccabe", "version": "0.5.9", "is_local": False},
|
||||
{
|
||||
"plugin": "pycodestyle",
|
||||
"version": "2.0.0",
|
||||
"is_local": False,
|
||||
},
|
||||
],
|
||||
),
|
||||
(
|
||||
[
|
||||
manager.PluginVersion("pycodestyle", "2.0.0", False),
|
||||
manager.PluginVersion("my-local", "0.0.1", True),
|
||||
manager.PluginVersion("mccabe", "0.5.9", False),
|
||||
],
|
||||
[
|
||||
{"plugin": "mccabe", "version": "0.5.9", "is_local": False},
|
||||
{"plugin": "my-local", "version": "0.0.1", "is_local": True},
|
||||
{
|
||||
"plugin": "pycodestyle",
|
||||
"version": "2.0.0",
|
||||
"is_local": False,
|
||||
},
|
||||
],
|
||||
),
|
||||
],
|
||||
)
|
||||
def test_plugins_from(plugins, expected):
|
||||
"""Test that we format plugins appropriately."""
|
||||
option_manager = mock.Mock(registered_plugins=set(plugins))
|
||||
assert expected == debug.plugins_from(option_manager)
|
||||
|
||||
|
||||
@mock.patch('platform.python_implementation', return_value='CPython')
|
||||
@mock.patch('platform.python_version', return_value='3.5.3')
|
||||
@mock.patch('platform.system', return_value='Linux')
|
||||
@mock.patch("platform.python_implementation", return_value="CPython")
|
||||
@mock.patch("platform.python_version", return_value="3.5.3")
|
||||
@mock.patch("platform.system", return_value="Linux")
|
||||
def test_information(system, pyversion, pyimpl):
|
||||
"""Verify that we return all the information we care about."""
|
||||
expected = {
|
||||
'version': '3.1.0',
|
||||
'plugins': [{'plugin': 'mccabe', 'version': '0.5.9',
|
||||
'is_local': False},
|
||||
{'plugin': 'pycodestyle', 'version': '2.0.0',
|
||||
'is_local': False}],
|
||||
'dependencies': [],
|
||||
'platform': {
|
||||
'python_implementation': 'CPython',
|
||||
'python_version': '3.5.3',
|
||||
'system': 'Linux',
|
||||
"version": "3.1.0",
|
||||
"plugins": [
|
||||
{"plugin": "mccabe", "version": "0.5.9", "is_local": False},
|
||||
{"plugin": "pycodestyle", "version": "2.0.0", "is_local": False},
|
||||
],
|
||||
"dependencies": [],
|
||||
"platform": {
|
||||
"python_implementation": "CPython",
|
||||
"python_version": "3.5.3",
|
||||
"system": "Linux",
|
||||
},
|
||||
}
|
||||
option_manager = mock.Mock(
|
||||
registered_plugins={
|
||||
manager.PluginVersion('pycodestyle', '2.0.0', False),
|
||||
manager.PluginVersion('mccabe', '0.5.9', False),
|
||||
manager.PluginVersion("pycodestyle", "2.0.0", False),
|
||||
manager.PluginVersion("mccabe", "0.5.9", False),
|
||||
},
|
||||
version='3.1.0',
|
||||
version="3.1.0",
|
||||
)
|
||||
assert expected == debug.information(option_manager)
|
||||
pyimpl.assert_called_once_with()
|
||||
|
|
@ -64,14 +95,16 @@ def test_information(system, pyversion, pyimpl):
|
|||
system.assert_called_once_with()
|
||||
|
||||
|
||||
@mock.patch('flake8.main.debug.print')
|
||||
@mock.patch('flake8.main.debug.information', return_value={})
|
||||
@mock.patch('json.dumps', return_value='{}')
|
||||
@mock.patch("flake8.main.debug.print")
|
||||
@mock.patch("flake8.main.debug.information", return_value={})
|
||||
@mock.patch("json.dumps", return_value="{}")
|
||||
def test_print_information_no_plugins(dumps, information, print_mock):
|
||||
"""Verify we print and exit only when we have plugins."""
|
||||
option_manager = mock.Mock(registered_plugins=set())
|
||||
action = debug.DebugAction(
|
||||
"--bug-report", dest="bug_report", option_manager=option_manager,
|
||||
"--bug-report",
|
||||
dest="bug_report",
|
||||
option_manager=option_manager,
|
||||
)
|
||||
assert action(None, None, None, None) is None
|
||||
assert dumps.called is False
|
||||
|
|
@ -79,21 +112,23 @@ def test_print_information_no_plugins(dumps, information, print_mock):
|
|||
assert print_mock.called is False
|
||||
|
||||
|
||||
@mock.patch('flake8.main.debug.print')
|
||||
@mock.patch('flake8.main.debug.information', return_value={})
|
||||
@mock.patch('json.dumps', return_value='{}')
|
||||
@mock.patch("flake8.main.debug.print")
|
||||
@mock.patch("flake8.main.debug.information", return_value={})
|
||||
@mock.patch("json.dumps", return_value="{}")
|
||||
def test_print_information(dumps, information, print_mock):
|
||||
"""Verify we print and exit only when we have plugins."""
|
||||
plugins = [
|
||||
manager.PluginVersion('pycodestyle', '2.0.0', False),
|
||||
manager.PluginVersion('mccabe', '0.5.9', False),
|
||||
manager.PluginVersion("pycodestyle", "2.0.0", False),
|
||||
manager.PluginVersion("mccabe", "0.5.9", False),
|
||||
]
|
||||
option_manager = mock.Mock(registered_plugins=set(plugins))
|
||||
action = debug.DebugAction(
|
||||
"--bug-report", dest="bug_report", option_manager=option_manager,
|
||||
"--bug-report",
|
||||
dest="bug_report",
|
||||
option_manager=option_manager,
|
||||
)
|
||||
with pytest.raises(SystemExit):
|
||||
action(None, None, None, None)
|
||||
print_mock.assert_called_once_with('{}')
|
||||
print_mock.assert_called_once_with("{}")
|
||||
dumps.assert_called_once_with({}, indent=2, sort_keys=True)
|
||||
information.assert_called_once_with(option_manager)
|
||||
|
|
|
|||
|
|
@ -7,21 +7,21 @@ from flake8 import exceptions
|
|||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'err',
|
||||
"err",
|
||||
(
|
||||
exceptions.FailedToLoadPlugin(
|
||||
plugin_name='plugin_name',
|
||||
exception=ValueError('boom!'),
|
||||
plugin_name="plugin_name",
|
||||
exception=ValueError("boom!"),
|
||||
),
|
||||
exceptions.InvalidSyntax(exception=ValueError('Unexpected token: $')),
|
||||
exceptions.InvalidSyntax(exception=ValueError("Unexpected token: $")),
|
||||
exceptions.PluginRequestedUnknownParameters(
|
||||
plugin={'plugin_name': 'plugin_name'},
|
||||
exception=ValueError('boom!'),
|
||||
plugin={"plugin_name": "plugin_name"},
|
||||
exception=ValueError("boom!"),
|
||||
),
|
||||
exceptions.PluginExecutionFailed(
|
||||
plugin={'plugin_name': 'plugin_name'},
|
||||
exception=ValueError('boom!'),
|
||||
)
|
||||
plugin={"plugin_name": "plugin_name"},
|
||||
exception=ValueError("boom!"),
|
||||
),
|
||||
),
|
||||
)
|
||||
def test_pickleable(err):
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ import flake8
|
|||
from flake8 import checker
|
||||
|
||||
|
||||
@mock.patch('flake8.processor.FileProcessor')
|
||||
@mock.patch("flake8.processor.FileProcessor")
|
||||
def test_run_ast_checks_handles_SyntaxErrors(FileProcessor): # noqa: N802,N803
|
||||
"""Stress our SyntaxError handling.
|
||||
|
||||
|
|
@ -15,26 +15,31 @@ def test_run_ast_checks_handles_SyntaxErrors(FileProcessor): # noqa: N802,N803
|
|||
"""
|
||||
processor = mock.Mock(lines=[])
|
||||
FileProcessor.return_value = processor
|
||||
processor.build_ast.side_effect = SyntaxError('Failed to build ast',
|
||||
('', 1, 5, 'foo(\n'))
|
||||
processor.build_ast.side_effect = SyntaxError(
|
||||
"Failed to build ast", ("", 1, 5, "foo(\n")
|
||||
)
|
||||
file_checker = checker.FileChecker(__file__, checks={}, options=object())
|
||||
|
||||
with mock.patch.object(file_checker, 'report') as report:
|
||||
with mock.patch.object(file_checker, "report") as report:
|
||||
file_checker.run_ast_checks()
|
||||
|
||||
report.assert_called_once_with(
|
||||
'E999', 1, 3,
|
||||
'SyntaxError: Failed to build ast',
|
||||
"E999",
|
||||
1,
|
||||
3,
|
||||
"SyntaxError: Failed to build ast",
|
||||
)
|
||||
|
||||
|
||||
@mock.patch('flake8.checker.FileChecker._make_processor', return_value=None)
|
||||
@mock.patch("flake8.checker.FileChecker._make_processor", return_value=None)
|
||||
def test_repr(*args):
|
||||
"""Verify we generate a correct repr."""
|
||||
file_checker = checker.FileChecker(
|
||||
'example.py', checks={}, options=object(),
|
||||
"example.py",
|
||||
checks={},
|
||||
options=object(),
|
||||
)
|
||||
assert repr(file_checker) == 'FileChecker for example.py'
|
||||
assert repr(file_checker) == "FileChecker for example.py"
|
||||
|
||||
|
||||
def test_nonexistent_file():
|
||||
|
|
@ -50,7 +55,7 @@ def test_nonexistent_file():
|
|||
|
||||
def test_raises_exception_on_failed_plugin(tmp_path, default_options):
|
||||
"""Checks that a failing plugin results in PluginExecutionFailed."""
|
||||
foobar = tmp_path / 'foobar.py'
|
||||
foobar = tmp_path / "foobar.py"
|
||||
foobar.write_text("I exist!") # Create temp file
|
||||
plugin = {
|
||||
"name": "failure",
|
||||
|
|
@ -60,6 +65,7 @@ def test_raises_exception_on_failed_plugin(tmp_path, default_options):
|
|||
}
|
||||
"""Verify a failing plugin results in an plugin error"""
|
||||
fchecker = checker.FileChecker(
|
||||
str(foobar), checks=[], options=default_options)
|
||||
str(foobar), checks=[], options=default_options
|
||||
)
|
||||
with pytest.raises(flake8.exceptions.PluginExecutionFailed):
|
||||
fchecker.run_check(plugin)
|
||||
|
|
|
|||
|
|
@ -17,7 +17,7 @@ def test_read_lines_splits_lines(default_options):
|
|||
|
||||
|
||||
def _lines_from_file(tmpdir, contents, options):
|
||||
f = tmpdir.join('f.py')
|
||||
f = tmpdir.join("f.py")
|
||||
# be careful to write the bytes exactly to avoid newline munging
|
||||
f.write_binary(contents)
|
||||
return processor.FileProcessor(f.strpath, options).lines
|
||||
|
|
@ -26,111 +26,125 @@ def _lines_from_file(tmpdir, contents, options):
|
|||
def test_read_lines_universal_newlines(tmpdir, default_options):
|
||||
r"""Verify that line endings are translated to \n."""
|
||||
lines = _lines_from_file(
|
||||
tmpdir, b'# coding: utf-8\r\nx = 1\r\n', default_options)
|
||||
assert lines == ['# coding: utf-8\n', 'x = 1\n']
|
||||
tmpdir, b"# coding: utf-8\r\nx = 1\r\n", default_options
|
||||
)
|
||||
assert lines == ["# coding: utf-8\n", "x = 1\n"]
|
||||
|
||||
|
||||
def test_read_lines_incorrect_utf_16(tmpdir, default_options):
|
||||
"""Verify that an incorrectly encoded file is read as latin-1."""
|
||||
lines = _lines_from_file(
|
||||
tmpdir, b'# coding: utf16\nx = 1\n', default_options)
|
||||
assert lines == ['# coding: utf16\n', 'x = 1\n']
|
||||
tmpdir, b"# coding: utf16\nx = 1\n", default_options
|
||||
)
|
||||
assert lines == ["# coding: utf16\n", "x = 1\n"]
|
||||
|
||||
|
||||
def test_read_lines_unknown_encoding(tmpdir, default_options):
|
||||
"""Verify that an unknown encoding is still read as latin-1."""
|
||||
lines = _lines_from_file(
|
||||
tmpdir, b'# coding: fake-encoding\nx = 1\n', default_options)
|
||||
assert lines == ['# coding: fake-encoding\n', 'x = 1\n']
|
||||
tmpdir, b"# coding: fake-encoding\nx = 1\n", default_options
|
||||
)
|
||||
assert lines == ["# coding: fake-encoding\n", "x = 1\n"]
|
||||
|
||||
|
||||
@pytest.mark.parametrize('first_line', [
|
||||
'\xEF\xBB\xBF"""Module docstring."""\n',
|
||||
'\uFEFF"""Module docstring."""\n',
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"first_line",
|
||||
[
|
||||
'\xEF\xBB\xBF"""Module docstring."""\n',
|
||||
'\uFEFF"""Module docstring."""\n',
|
||||
],
|
||||
)
|
||||
def test_strip_utf_bom(first_line, default_options):
|
||||
r"""Verify that we strip '\xEF\xBB\xBF' from the first line."""
|
||||
lines = [first_line]
|
||||
file_processor = processor.FileProcessor('-', default_options, lines[:])
|
||||
file_processor = processor.FileProcessor("-", default_options, lines[:])
|
||||
assert file_processor.lines != lines
|
||||
assert file_processor.lines[0] == '"""Module docstring."""\n'
|
||||
|
||||
|
||||
@pytest.mark.parametrize('lines, expected', [
|
||||
(['\xEF\xBB\xBF"""Module docstring."""\n'], False),
|
||||
(['\uFEFF"""Module docstring."""\n'], False),
|
||||
(['#!/usr/bin/python', '# flake8 is great', 'a = 1'], False),
|
||||
(['#!/usr/bin/python', '# flake8: noqa', 'a = 1'], True),
|
||||
(['#!/usr/bin/python', '# flake8:noqa', 'a = 1'], True),
|
||||
(['# flake8: noqa', '#!/usr/bin/python', 'a = 1'], True),
|
||||
(['# flake8:noqa', '#!/usr/bin/python', 'a = 1'], True),
|
||||
(['#!/usr/bin/python', 'a = 1', '# flake8: noqa'], True),
|
||||
(['#!/usr/bin/python', 'a = 1', '# flake8:noqa'], True),
|
||||
(['#!/usr/bin/python', 'a = 1 # flake8: noqa'], False),
|
||||
(['#!/usr/bin/python', 'a = 1 # flake8:noqa'], False),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"lines, expected",
|
||||
[
|
||||
(['\xEF\xBB\xBF"""Module docstring."""\n'], False),
|
||||
(['\uFEFF"""Module docstring."""\n'], False),
|
||||
(["#!/usr/bin/python", "# flake8 is great", "a = 1"], False),
|
||||
(["#!/usr/bin/python", "# flake8: noqa", "a = 1"], True),
|
||||
(["#!/usr/bin/python", "# flake8:noqa", "a = 1"], True),
|
||||
(["# flake8: noqa", "#!/usr/bin/python", "a = 1"], True),
|
||||
(["# flake8:noqa", "#!/usr/bin/python", "a = 1"], True),
|
||||
(["#!/usr/bin/python", "a = 1", "# flake8: noqa"], True),
|
||||
(["#!/usr/bin/python", "a = 1", "# flake8:noqa"], True),
|
||||
(["#!/usr/bin/python", "a = 1 # flake8: noqa"], False),
|
||||
(["#!/usr/bin/python", "a = 1 # flake8:noqa"], False),
|
||||
],
|
||||
)
|
||||
def test_should_ignore_file(lines, expected, default_options):
|
||||
"""Verify that we ignore a file if told to."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines)
|
||||
file_processor = processor.FileProcessor("-", default_options, lines)
|
||||
assert file_processor.should_ignore_file() is expected
|
||||
|
||||
|
||||
def test_should_ignore_file_to_handle_disable_noqa(default_options):
|
||||
"""Verify that we ignore a file if told to."""
|
||||
lines = ['# flake8: noqa']
|
||||
file_processor = processor.FileProcessor('-', default_options, lines)
|
||||
lines = ["# flake8: noqa"]
|
||||
file_processor = processor.FileProcessor("-", default_options, lines)
|
||||
assert file_processor.should_ignore_file() is True
|
||||
default_options.disable_noqa = True
|
||||
file_processor = processor.FileProcessor('-', default_options, lines)
|
||||
file_processor = processor.FileProcessor("-", default_options, lines)
|
||||
assert file_processor.should_ignore_file() is False
|
||||
|
||||
|
||||
@mock.patch('flake8.utils.stdin_get_value')
|
||||
@mock.patch("flake8.utils.stdin_get_value")
|
||||
def test_read_lines_from_stdin(stdin_get_value, default_options):
|
||||
"""Verify that we use our own utility function to retrieve stdin."""
|
||||
stdin_get_value.return_value = ''
|
||||
processor.FileProcessor('-', default_options)
|
||||
stdin_get_value.return_value = ""
|
||||
processor.FileProcessor("-", default_options)
|
||||
stdin_get_value.assert_called_once_with()
|
||||
|
||||
|
||||
@mock.patch('flake8.utils.stdin_get_value')
|
||||
@mock.patch("flake8.utils.stdin_get_value")
|
||||
def test_stdin_filename_attribute(stdin_get_value, default_options):
|
||||
"""Verify that we update the filename attribute."""
|
||||
stdin_get_value.return_value = ''
|
||||
file_processor = processor.FileProcessor('-', default_options)
|
||||
assert file_processor.filename == 'stdin'
|
||||
stdin_get_value.return_value = ""
|
||||
file_processor = processor.FileProcessor("-", default_options)
|
||||
assert file_processor.filename == "stdin"
|
||||
|
||||
|
||||
@mock.patch('flake8.utils.stdin_get_value')
|
||||
@mock.patch("flake8.utils.stdin_get_value")
|
||||
def test_read_lines_uses_display_name(stdin_get_value, default_options):
|
||||
"""Verify that when processing stdin we use a display name if present."""
|
||||
default_options.stdin_display_name = 'display_name.py'
|
||||
stdin_get_value.return_value = ''
|
||||
file_processor = processor.FileProcessor('-', default_options)
|
||||
assert file_processor.filename == 'display_name.py'
|
||||
default_options.stdin_display_name = "display_name.py"
|
||||
stdin_get_value.return_value = ""
|
||||
file_processor = processor.FileProcessor("-", default_options)
|
||||
assert file_processor.filename == "display_name.py"
|
||||
|
||||
|
||||
@mock.patch('flake8.utils.stdin_get_value')
|
||||
@mock.patch("flake8.utils.stdin_get_value")
|
||||
def test_read_lines_ignores_empty_display_name(
|
||||
stdin_get_value, default_options,
|
||||
stdin_get_value,
|
||||
default_options,
|
||||
):
|
||||
"""Verify that when processing stdin we use a display name if present."""
|
||||
stdin_get_value.return_value = ''
|
||||
default_options.stdin_display_name = ''
|
||||
file_processor = processor.FileProcessor('-', default_options)
|
||||
assert file_processor.filename == 'stdin'
|
||||
stdin_get_value.return_value = ""
|
||||
default_options.stdin_display_name = ""
|
||||
file_processor = processor.FileProcessor("-", default_options)
|
||||
assert file_processor.filename == "stdin"
|
||||
|
||||
|
||||
def test_noqa_line_for(default_options):
|
||||
"""Verify we grab the correct line from the cached lines."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'Line 1\n',
|
||||
'Line 2\n',
|
||||
'Line 3\n',
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-",
|
||||
default_options,
|
||||
lines=[
|
||||
"Line 1\n",
|
||||
"Line 2\n",
|
||||
"Line 3\n",
|
||||
],
|
||||
)
|
||||
|
||||
for i in range(1, 4):
|
||||
assert file_processor.noqa_line_for(i) == f'Line {i}\n'
|
||||
assert file_processor.noqa_line_for(i) == f"Line {i}\n"
|
||||
|
||||
|
||||
def test_noqa_line_for_continuation(default_options):
|
||||
|
|
@ -145,15 +159,15 @@ world
|
|||
""" # 7
|
||||
'''
|
||||
lines = src.splitlines(True)
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=lines)
|
||||
file_processor = processor.FileProcessor("-", default_options, lines=lines)
|
||||
|
||||
assert file_processor.noqa_line_for(0) is None
|
||||
|
||||
l_1_2 = 'from foo \\\n import bar # 2\n'
|
||||
l_1_2 = "from foo \\\n import bar # 2\n"
|
||||
assert file_processor.noqa_line_for(1) == l_1_2
|
||||
assert file_processor.noqa_line_for(2) == l_1_2
|
||||
|
||||
assert file_processor.noqa_line_for(3) == '\n'
|
||||
assert file_processor.noqa_line_for(3) == "\n"
|
||||
|
||||
l_4_7 = 'x = """\nhello\nworld\n""" # 7\n'
|
||||
for i in (4, 5, 6, 7):
|
||||
|
|
@ -164,76 +178,104 @@ world
|
|||
|
||||
def test_noqa_line_for_no_eol_at_end_of_file(default_options):
|
||||
"""Verify that we properly handle noqa line at the end of the file."""
|
||||
src = 'from foo \\\nimport bar' # no end of file newline
|
||||
src = "from foo \\\nimport bar" # no end of file newline
|
||||
lines = src.splitlines(True)
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=lines)
|
||||
file_processor = processor.FileProcessor("-", default_options, lines=lines)
|
||||
|
||||
l_1_2 = 'from foo \\\nimport bar'
|
||||
l_1_2 = "from foo \\\nimport bar"
|
||||
assert file_processor.noqa_line_for(1) == l_1_2
|
||||
assert file_processor.noqa_line_for(2) == l_1_2
|
||||
|
||||
|
||||
def test_next_line(default_options):
|
||||
"""Verify we update the file_processor state for each new line."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'Line 1',
|
||||
'Line 2',
|
||||
'Line 3',
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-",
|
||||
default_options,
|
||||
lines=[
|
||||
"Line 1",
|
||||
"Line 2",
|
||||
"Line 3",
|
||||
],
|
||||
)
|
||||
|
||||
for i in range(1, 4):
|
||||
assert file_processor.next_line() == f'Line {i}'
|
||||
assert file_processor.next_line() == f"Line {i}"
|
||||
assert file_processor.line_number == i
|
||||
|
||||
|
||||
@pytest.mark.parametrize('params, args, expected_kwargs', [
|
||||
({'blank_before': True, 'blank_lines': True},
|
||||
None,
|
||||
{'blank_before': 0, 'blank_lines': 0}),
|
||||
({'noqa': True, 'fake': True},
|
||||
{'fake': 'foo'},
|
||||
{'noqa': False, 'fake': 'foo'}),
|
||||
({'blank_before': True, 'blank_lines': True, 'noqa': True},
|
||||
{'blank_before': 10, 'blank_lines': 5, 'noqa': True},
|
||||
{'blank_before': 10, 'blank_lines': 5, 'noqa': True}),
|
||||
({}, {'fake': 'foo'}, {'fake': 'foo'}),
|
||||
({'non-existent': False}, {'fake': 'foo'}, {'fake': 'foo'}),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"params, args, expected_kwargs",
|
||||
[
|
||||
(
|
||||
{"blank_before": True, "blank_lines": True},
|
||||
None,
|
||||
{"blank_before": 0, "blank_lines": 0},
|
||||
),
|
||||
(
|
||||
{"noqa": True, "fake": True},
|
||||
{"fake": "foo"},
|
||||
{"noqa": False, "fake": "foo"},
|
||||
),
|
||||
(
|
||||
{"blank_before": True, "blank_lines": True, "noqa": True},
|
||||
{"blank_before": 10, "blank_lines": 5, "noqa": True},
|
||||
{"blank_before": 10, "blank_lines": 5, "noqa": True},
|
||||
),
|
||||
({}, {"fake": "foo"}, {"fake": "foo"}),
|
||||
({"non-existent": False}, {"fake": "foo"}, {"fake": "foo"}),
|
||||
],
|
||||
)
|
||||
def test_keyword_arguments_for(params, args, expected_kwargs, default_options):
|
||||
"""Verify the keyword args are generated properly."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'Line 1',
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-",
|
||||
default_options,
|
||||
lines=[
|
||||
"Line 1",
|
||||
],
|
||||
)
|
||||
kwargs_for = file_processor.keyword_arguments_for
|
||||
|
||||
assert kwargs_for(params, args) == expected_kwargs
|
||||
|
||||
|
||||
def test_keyword_arguments_for_does_not_handle_attribute_errors(
|
||||
default_options,
|
||||
default_options,
|
||||
):
|
||||
"""Verify we re-raise AttributeErrors."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'Line 1',
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-",
|
||||
default_options,
|
||||
lines=[
|
||||
"Line 1",
|
||||
],
|
||||
)
|
||||
|
||||
with pytest.raises(AttributeError):
|
||||
file_processor.keyword_arguments_for({'fake': True})
|
||||
file_processor.keyword_arguments_for({"fake": True})
|
||||
|
||||
|
||||
@pytest.mark.parametrize('unsplit_line, expected_lines', [
|
||||
('line', []),
|
||||
('line 1\n', ['line 1']),
|
||||
('line 1\nline 2\n', ['line 1', 'line 2']),
|
||||
('line 1\n\nline 2\n', ['line 1', '', 'line 2']),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"unsplit_line, expected_lines",
|
||||
[
|
||||
("line", []),
|
||||
("line 1\n", ["line 1"]),
|
||||
("line 1\nline 2\n", ["line 1", "line 2"]),
|
||||
("line 1\n\nline 2\n", ["line 1", "", "line 2"]),
|
||||
],
|
||||
)
|
||||
def test_split_line(unsplit_line, expected_lines, default_options):
|
||||
"""Verify the token line splitting."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'Line 1',
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-",
|
||||
default_options,
|
||||
lines=[
|
||||
"Line 1",
|
||||
],
|
||||
)
|
||||
|
||||
token = (1, unsplit_line, (0, 0), (0, 0), '')
|
||||
token = (1, unsplit_line, (0, 0), (0, 0), "")
|
||||
actual_lines = list(file_processor.split_line(token))
|
||||
assert expected_lines == actual_lines
|
||||
|
||||
|
|
@ -242,9 +284,9 @@ def test_split_line(unsplit_line, expected_lines, default_options):
|
|||
|
||||
def test_build_ast(default_options):
|
||||
"""Verify the logic for how we build an AST for plugins."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'a = 1\n'
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-", default_options, lines=["a = 1\n"]
|
||||
)
|
||||
|
||||
module = file_processor.build_ast()
|
||||
assert isinstance(module, ast.Module)
|
||||
|
|
@ -252,25 +294,25 @@ def test_build_ast(default_options):
|
|||
|
||||
def test_next_logical_line_updates_the_previous_logical_line(default_options):
|
||||
"""Verify that we update our tracking of the previous logical line."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'a = 1\n'
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-", default_options, lines=["a = 1\n"]
|
||||
)
|
||||
|
||||
file_processor.indent_level = 1
|
||||
file_processor.logical_line = 'a = 1'
|
||||
assert file_processor.previous_logical == ''
|
||||
file_processor.logical_line = "a = 1"
|
||||
assert file_processor.previous_logical == ""
|
||||
assert file_processor.previous_indent_level == 0
|
||||
|
||||
file_processor.next_logical_line()
|
||||
assert file_processor.previous_logical == 'a = 1'
|
||||
assert file_processor.previous_logical == "a = 1"
|
||||
assert file_processor.previous_indent_level == 1
|
||||
|
||||
|
||||
def test_visited_new_blank_line(default_options):
|
||||
"""Verify we update the number of blank lines seen."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'a = 1\n'
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-", default_options, lines=["a = 1\n"]
|
||||
)
|
||||
|
||||
assert file_processor.blank_lines == 0
|
||||
file_processor.visited_new_blank_line()
|
||||
|
|
@ -279,9 +321,9 @@ def test_visited_new_blank_line(default_options):
|
|||
|
||||
def test_inside_multiline(default_options):
|
||||
"""Verify we update the line number and reset multiline."""
|
||||
file_processor = processor.FileProcessor('-', default_options, lines=[
|
||||
'a = 1\n'
|
||||
])
|
||||
file_processor = processor.FileProcessor(
|
||||
"-", default_options, lines=["a = 1\n"]
|
||||
)
|
||||
|
||||
assert file_processor.multiline is False
|
||||
assert file_processor.line_number == 0
|
||||
|
|
@ -292,63 +334,87 @@ def test_inside_multiline(default_options):
|
|||
assert file_processor.multiline is False
|
||||
|
||||
|
||||
@pytest.mark.parametrize('string, expected', [
|
||||
('""', '""'),
|
||||
("''", "''"),
|
||||
('"a"', '"x"'),
|
||||
("'a'", "'x'"),
|
||||
('"x"', '"x"'),
|
||||
("'x'", "'x'"),
|
||||
('"abcdef"', '"xxxxxx"'),
|
||||
("'abcdef'", "'xxxxxx'"),
|
||||
('""""""', '""""""'),
|
||||
("''''''", "''''''"),
|
||||
('"""a"""', '"""x"""'),
|
||||
("'''a'''", "'''x'''"),
|
||||
('"""x"""', '"""x"""'),
|
||||
("'''x'''", "'''x'''"),
|
||||
('"""abcdef"""', '"""xxxxxx"""'),
|
||||
("'''abcdef'''", "'''xxxxxx'''"),
|
||||
('"""xxxxxx"""', '"""xxxxxx"""'),
|
||||
("'''xxxxxx'''", "'''xxxxxx'''"),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"string, expected",
|
||||
[
|
||||
('""', '""'),
|
||||
("''", "''"),
|
||||
('"a"', '"x"'),
|
||||
("'a'", "'x'"),
|
||||
('"x"', '"x"'),
|
||||
("'x'", "'x'"),
|
||||
('"abcdef"', '"xxxxxx"'),
|
||||
("'abcdef'", "'xxxxxx'"),
|
||||
('""""""', '""""""'),
|
||||
("''''''", "''''''"),
|
||||
('"""a"""', '"""x"""'),
|
||||
("'''a'''", "'''x'''"),
|
||||
('"""x"""', '"""x"""'),
|
||||
("'''x'''", "'''x'''"),
|
||||
('"""abcdef"""', '"""xxxxxx"""'),
|
||||
("'''abcdef'''", "'''xxxxxx'''"),
|
||||
('"""xxxxxx"""', '"""xxxxxx"""'),
|
||||
("'''xxxxxx'''", "'''xxxxxx'''"),
|
||||
],
|
||||
)
|
||||
def test_mutate_string(string, expected, default_options):
|
||||
"""Verify we appropriately mutate the string to sanitize it."""
|
||||
actual = processor.mutate_string(string)
|
||||
assert expected == actual
|
||||
|
||||
|
||||
@pytest.mark.parametrize('string, expected', [
|
||||
(' ', 4),
|
||||
(' ', 6),
|
||||
('\t', 8),
|
||||
('\t\t', 16),
|
||||
(' \t', 8),
|
||||
(' \t', 16),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"string, expected",
|
||||
[
|
||||
(" ", 4),
|
||||
(" ", 6),
|
||||
("\t", 8),
|
||||
("\t\t", 16),
|
||||
(" \t", 8),
|
||||
(" \t", 16),
|
||||
],
|
||||
)
|
||||
def test_expand_indent(string, expected):
|
||||
"""Verify we correctly measure the amount of indentation."""
|
||||
actual = processor.expand_indent(string)
|
||||
assert expected == actual
|
||||
|
||||
|
||||
@pytest.mark.parametrize('token, log_string', [
|
||||
[(tokenize.COMMENT, '# this is a comment',
|
||||
(1, 0), # (start_row, start_column)
|
||||
(1, 19), # (end_ro, end_column)
|
||||
'# this is a comment',),
|
||||
"l.1\t[:19]\tCOMMENT\t'# this is a comment'"],
|
||||
[(tokenize.COMMENT, '# this is a comment',
|
||||
(1, 5), # (start_row, start_column)
|
||||
(1, 19), # (end_ro, end_column)
|
||||
'# this is a comment',),
|
||||
"l.1\t[5:19]\tCOMMENT\t'# this is a comment'"],
|
||||
[(tokenize.COMMENT, '# this is a comment',
|
||||
(1, 0), # (start_row, start_column)
|
||||
(2, 19), # (end_ro, end_column)
|
||||
'# this is a comment',),
|
||||
"l.1\tl.2\tCOMMENT\t'# this is a comment'"],
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"token, log_string",
|
||||
[
|
||||
[
|
||||
(
|
||||
tokenize.COMMENT,
|
||||
"# this is a comment",
|
||||
(1, 0), # (start_row, start_column)
|
||||
(1, 19), # (end_ro, end_column)
|
||||
"# this is a comment",
|
||||
),
|
||||
"l.1\t[:19]\tCOMMENT\t'# this is a comment'",
|
||||
],
|
||||
[
|
||||
(
|
||||
tokenize.COMMENT,
|
||||
"# this is a comment",
|
||||
(1, 5), # (start_row, start_column)
|
||||
(1, 19), # (end_ro, end_column)
|
||||
"# this is a comment",
|
||||
),
|
||||
"l.1\t[5:19]\tCOMMENT\t'# this is a comment'",
|
||||
],
|
||||
[
|
||||
(
|
||||
tokenize.COMMENT,
|
||||
"# this is a comment",
|
||||
(1, 0), # (start_row, start_column)
|
||||
(2, 19), # (end_ro, end_column)
|
||||
"# this is a comment",
|
||||
),
|
||||
"l.1\tl.2\tCOMMENT\t'# this is a comment'",
|
||||
],
|
||||
],
|
||||
)
|
||||
def test_log_token(token, log_string):
|
||||
"""Verify we use the log object passed in."""
|
||||
log = mock.Mock()
|
||||
|
|
@ -359,15 +425,18 @@ def test_log_token(token, log_string):
|
|||
)
|
||||
|
||||
|
||||
@pytest.mark.parametrize('current_count, token_text, expected', [
|
||||
(0, '(', 1),
|
||||
(0, '[', 1),
|
||||
(0, '{', 1),
|
||||
(1, ')', 0),
|
||||
(1, ']', 0),
|
||||
(1, '}', 0),
|
||||
(10, '+', 10),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"current_count, token_text, expected",
|
||||
[
|
||||
(0, "(", 1),
|
||||
(0, "[", 1),
|
||||
(0, "{", 1),
|
||||
(1, ")", 0),
|
||||
(1, "]", 0),
|
||||
(1, "}", 0),
|
||||
(10, "+", 10),
|
||||
],
|
||||
)
|
||||
def test_count_parentheses(current_count, token_text, expected):
|
||||
"""Verify our arithmetic is correct."""
|
||||
assert processor.count_parentheses(current_count, token_text) == expected
|
||||
|
|
|
|||
|
|
@ -7,8 +7,8 @@ from flake8.formatting import default
|
|||
|
||||
def options(**kwargs):
|
||||
"""Create an argparse.Namespace instance."""
|
||||
kwargs.setdefault('output_file', None)
|
||||
kwargs.setdefault('tee', False)
|
||||
kwargs.setdefault("output_file", None)
|
||||
kwargs.setdefault("tee", False)
|
||||
return argparse.Namespace(**kwargs)
|
||||
|
||||
|
||||
|
|
@ -18,22 +18,23 @@ def test_caches_filenames_already_printed():
|
|||
assert formatter.filenames_already_printed == set()
|
||||
|
||||
formatter.format(
|
||||
style_guide.Violation('code', 'file.py', 1, 1, 'text', 'l'))
|
||||
assert formatter.filenames_already_printed == {'file.py'}
|
||||
style_guide.Violation("code", "file.py", 1, 1, "text", "l")
|
||||
)
|
||||
assert formatter.filenames_already_printed == {"file.py"}
|
||||
|
||||
|
||||
def test_only_returns_a_string_once_from_format():
|
||||
"""Verify format ignores the second error with the same filename."""
|
||||
formatter = default.FilenameOnly(options())
|
||||
error = style_guide.Violation('code', 'file.py', 1, 1, 'text', '1')
|
||||
error = style_guide.Violation("code", "file.py", 1, 1, "text", "1")
|
||||
|
||||
assert formatter.format(error) == 'file.py'
|
||||
assert formatter.format(error) == "file.py"
|
||||
assert formatter.format(error) is None
|
||||
|
||||
|
||||
def test_show_source_returns_nothing():
|
||||
"""Verify show_source returns nothing."""
|
||||
formatter = default.FilenameOnly(options())
|
||||
error = style_guide.Violation('code', 'file.py', 1, 1, 'text', '1')
|
||||
error = style_guide.Violation("code", "file.py", 1, 1, "text", "1")
|
||||
|
||||
assert formatter.show_source(error) is None
|
||||
|
|
|
|||
|
|
@ -23,8 +23,8 @@ def test_get_local_plugins_uses_cli_config():
|
|||
config_finder = mock.MagicMock()
|
||||
config_finder.cli_config.return_value = config_obj
|
||||
config_finder.ignore_config_files = False
|
||||
config_obj.get.return_value = ''
|
||||
config_file_value = 'foo.ini'
|
||||
config_obj.get.return_value = ""
|
||||
config_file_value = "foo.ini"
|
||||
config_finder.config_file = config_file_value
|
||||
|
||||
config.get_local_plugins(config_finder)
|
||||
|
|
@ -34,12 +34,12 @@ def test_get_local_plugins_uses_cli_config():
|
|||
|
||||
def test_get_local_plugins():
|
||||
"""Verify get_local_plugins returns expected plugins."""
|
||||
config_fixture_path = 'tests/fixtures/config_files/local-plugin.ini'
|
||||
config_finder = config.ConfigFileFinder('flake8')
|
||||
config_fixture_path = "tests/fixtures/config_files/local-plugin.ini"
|
||||
config_finder = config.ConfigFileFinder("flake8")
|
||||
|
||||
with mock.patch.object(config_finder, 'local_config_files') as localcfs:
|
||||
with mock.patch.object(config_finder, "local_config_files") as localcfs:
|
||||
localcfs.return_value = [config_fixture_path]
|
||||
local_plugins = config.get_local_plugins(config_finder)
|
||||
|
||||
assert local_plugins.extension == ['XE = test_plugins:ExtensionTestPlugin']
|
||||
assert local_plugins.report == ['XR = test_plugins:ReportTestPlugin']
|
||||
assert local_plugins.extension == ["XE = test_plugins:ExtensionTestPlugin"]
|
||||
assert local_plugins.report == ["XR = test_plugins:ReportTestPlugin"]
|
||||
|
|
|
|||
|
|
@ -21,12 +21,14 @@ def test_get_style_guide():
|
|||
)
|
||||
mockedapp = mock.Mock()
|
||||
mockedapp.parse_preliminary_options.return_value = (prelim_opts, [])
|
||||
mockedapp.program = 'flake8'
|
||||
with mock.patch('flake8.api.legacy.config.ConfigFileFinder') as mock_config_finder: # noqa: E501
|
||||
mockedapp.program = "flake8"
|
||||
with mock.patch(
|
||||
"flake8.api.legacy.config.ConfigFileFinder"
|
||||
) as mock_config_finder: # noqa: E501
|
||||
config_finder = ConfigFileFinder(mockedapp.program)
|
||||
mock_config_finder.return_value = config_finder
|
||||
|
||||
with mock.patch('flake8.main.application.Application') as application:
|
||||
with mock.patch("flake8.main.application.Application") as application:
|
||||
application.return_value = mockedapp
|
||||
style_guide = api.get_style_guide()
|
||||
|
||||
|
|
@ -35,7 +37,8 @@ def test_get_style_guide():
|
|||
mockedapp.find_plugins.assert_called_once_with(config_finder)
|
||||
mockedapp.register_plugin_options.assert_called_once_with()
|
||||
mockedapp.parse_configuration_and_cli.assert_called_once_with(
|
||||
config_finder, [])
|
||||
config_finder, []
|
||||
)
|
||||
mockedapp.make_formatter.assert_called_once_with()
|
||||
mockedapp.make_guide.assert_called_once_with()
|
||||
mockedapp.make_file_checker_manager.assert_called_once_with()
|
||||
|
|
@ -45,22 +48,22 @@ def test_get_style_guide():
|
|||
def test_styleguide_options():
|
||||
"""Show that we proxy the StyleGuide.options attribute."""
|
||||
app = mock.Mock()
|
||||
app.options = 'options'
|
||||
app.options = "options"
|
||||
style_guide = api.StyleGuide(app)
|
||||
assert style_guide.options == 'options'
|
||||
assert style_guide.options == "options"
|
||||
|
||||
|
||||
def test_styleguide_paths():
|
||||
"""Show that we proxy the StyleGuide.paths attribute."""
|
||||
app = mock.Mock()
|
||||
app.paths = 'paths'
|
||||
app.paths = "paths"
|
||||
style_guide = api.StyleGuide(app)
|
||||
assert style_guide.paths == 'paths'
|
||||
assert style_guide.paths == "paths"
|
||||
|
||||
|
||||
def test_styleguide_check_files():
|
||||
"""Verify we call the right application methods."""
|
||||
paths = ['foo', 'bar']
|
||||
paths = ["foo", "bar"]
|
||||
app = mock.Mock()
|
||||
style_guide = api.StyleGuide(app)
|
||||
report = style_guide.check_files(paths)
|
||||
|
|
@ -80,8 +83,8 @@ def test_styleguide_excluded():
|
|||
file_checker_manager = app.file_checker_manager = mock.Mock()
|
||||
style_guide = api.StyleGuide(app)
|
||||
|
||||
style_guide.excluded('file.py')
|
||||
file_checker_manager.is_path_excluded.assert_called_once_with('file.py')
|
||||
style_guide.excluded("file.py")
|
||||
file_checker_manager.is_path_excluded.assert_called_once_with("file.py")
|
||||
|
||||
|
||||
def test_styleguide_excluded_with_parent():
|
||||
|
|
@ -95,10 +98,10 @@ def test_styleguide_excluded_with_parent():
|
|||
file_checker_manager.is_path_excluded.return_value = False
|
||||
style_guide = api.StyleGuide(app)
|
||||
|
||||
style_guide.excluded('file.py', 'parent')
|
||||
style_guide.excluded("file.py", "parent")
|
||||
assert file_checker_manager.is_path_excluded.call_args_list == [
|
||||
mock.call('file.py'),
|
||||
mock.call(os.path.join('parent', 'file.py')),
|
||||
mock.call("file.py"),
|
||||
mock.call(os.path.join("parent", "file.py")),
|
||||
]
|
||||
|
||||
|
||||
|
|
@ -123,7 +126,7 @@ def test_styleguide_init_report_with_non_subclass():
|
|||
|
||||
def test_styleguide_init_report():
|
||||
"""Verify we do the right incantation for the Application."""
|
||||
app = mock.Mock(guide='fake')
|
||||
app = mock.Mock(guide="fake")
|
||||
style_guide = api.StyleGuide(app)
|
||||
|
||||
class FakeFormatter(formatter.BaseFormatter):
|
||||
|
|
@ -140,16 +143,16 @@ def test_styleguide_input_file():
|
|||
"""Verify we call StyleGuide.check_files with the filename."""
|
||||
app = mock.Mock()
|
||||
style_guide = api.StyleGuide(app)
|
||||
with mock.patch.object(style_guide, 'check_files') as check_files:
|
||||
style_guide.input_file('file.py')
|
||||
check_files.assert_called_once_with(['file.py'])
|
||||
with mock.patch.object(style_guide, "check_files") as check_files:
|
||||
style_guide.input_file("file.py")
|
||||
check_files.assert_called_once_with(["file.py"])
|
||||
|
||||
|
||||
def test_report_total_errors():
|
||||
"""Verify total errors is just a proxy attribute."""
|
||||
app = mock.Mock(result_count='Fake count')
|
||||
app = mock.Mock(result_count="Fake count")
|
||||
report = api.Report(app)
|
||||
assert report.total_errors == 'Fake count'
|
||||
assert report.total_errors == "Fake count"
|
||||
|
||||
|
||||
def test_report_get_statistics():
|
||||
|
|
@ -160,5 +163,5 @@ def test_report_get_statistics():
|
|||
app = mock.Mock(guide=style_guide)
|
||||
|
||||
report = api.Report(app)
|
||||
assert report.get_statistics('E') == []
|
||||
stats.statistics_for.assert_called_once_with('E')
|
||||
assert report.get_statistics("E") == []
|
||||
stats.statistics_for.assert_called_once_with("E")
|
||||
|
|
|
|||
|
|
@ -11,47 +11,54 @@ from flake8.options import manager
|
|||
@pytest.fixture
|
||||
def optmanager():
|
||||
"""Generate an OptionManager with simple values."""
|
||||
return manager.OptionManager(prog='flake8', version='3.0.0a1')
|
||||
return manager.OptionManager(prog="flake8", version="3.0.0a1")
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def config_finder():
|
||||
"""Generate a simple ConfigFileFinder."""
|
||||
return config.ConfigFileFinder('flake8')
|
||||
return config.ConfigFileFinder("flake8")
|
||||
|
||||
|
||||
def test_parse_cli_config(optmanager, config_finder):
|
||||
"""Parse the specified config file as a cli config file."""
|
||||
optmanager.add_option('--exclude', parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True)
|
||||
optmanager.add_option('--ignore', parse_from_config=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option('--quiet', parse_from_config=True,
|
||||
action='count')
|
||||
optmanager.add_option(
|
||||
"--exclude",
|
||||
parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True,
|
||||
)
|
||||
optmanager.add_option(
|
||||
"--ignore", parse_from_config=True, comma_separated_list=True
|
||||
)
|
||||
optmanager.add_option("--quiet", parse_from_config=True, action="count")
|
||||
parser = config.MergedConfigParser(optmanager, config_finder)
|
||||
|
||||
config_file = 'tests/fixtures/config_files/cli-specified.ini'
|
||||
config_file = "tests/fixtures/config_files/cli-specified.ini"
|
||||
parsed_config = parser.parse_cli_config(config_file)
|
||||
|
||||
config_dir = os.path.dirname(config_file)
|
||||
assert parsed_config == {
|
||||
'ignore': ['E123', 'W234', 'E111'],
|
||||
'exclude': [
|
||||
os.path.abspath(os.path.join(config_dir, 'foo/')),
|
||||
os.path.abspath(os.path.join(config_dir, 'bar/')),
|
||||
os.path.abspath(os.path.join(config_dir, 'bogus/')),
|
||||
"ignore": ["E123", "W234", "E111"],
|
||||
"exclude": [
|
||||
os.path.abspath(os.path.join(config_dir, "foo/")),
|
||||
os.path.abspath(os.path.join(config_dir, "bar/")),
|
||||
os.path.abspath(os.path.join(config_dir, "bogus/")),
|
||||
],
|
||||
'quiet': 1,
|
||||
"quiet": 1,
|
||||
}
|
||||
|
||||
|
||||
@pytest.mark.parametrize('filename,is_configured_by', [
|
||||
('tests/fixtures/config_files/cli-specified.ini', True),
|
||||
('tests/fixtures/config_files/no-flake8-section.ini', False),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"filename,is_configured_by",
|
||||
[
|
||||
("tests/fixtures/config_files/cli-specified.ini", True),
|
||||
("tests/fixtures/config_files/no-flake8-section.ini", False),
|
||||
],
|
||||
)
|
||||
def test_is_configured_by(
|
||||
filename, is_configured_by, optmanager, config_finder):
|
||||
filename, is_configured_by, optmanager, config_finder
|
||||
):
|
||||
"""Verify the behaviour of the is_configured_by method."""
|
||||
parsed_config, _ = config.ConfigFileFinder._read_config(filename)
|
||||
parser = config.MergedConfigParser(optmanager, config_finder)
|
||||
|
|
@ -61,83 +68,94 @@ def test_is_configured_by(
|
|||
|
||||
def test_parse_user_config(optmanager, config_finder):
|
||||
"""Verify parsing of user config files."""
|
||||
optmanager.add_option('--exclude', parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True)
|
||||
optmanager.add_option('--ignore', parse_from_config=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option('--quiet', parse_from_config=True,
|
||||
action='count')
|
||||
optmanager.add_option(
|
||||
"--exclude",
|
||||
parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True,
|
||||
)
|
||||
optmanager.add_option(
|
||||
"--ignore", parse_from_config=True, comma_separated_list=True
|
||||
)
|
||||
optmanager.add_option("--quiet", parse_from_config=True, action="count")
|
||||
parser = config.MergedConfigParser(optmanager, config_finder)
|
||||
|
||||
config_finder.user_config_file = ('tests/fixtures/config_files/'
|
||||
'cli-specified.ini')
|
||||
config_finder.user_config_file = (
|
||||
"tests/fixtures/config_files/" "cli-specified.ini"
|
||||
)
|
||||
parsed_config = parser.parse_user_config()
|
||||
|
||||
assert parsed_config == {
|
||||
'ignore': ['E123', 'W234', 'E111'],
|
||||
'exclude': [
|
||||
os.path.abspath('foo/'),
|
||||
os.path.abspath('bar/'),
|
||||
os.path.abspath('bogus/'),
|
||||
"ignore": ["E123", "W234", "E111"],
|
||||
"exclude": [
|
||||
os.path.abspath("foo/"),
|
||||
os.path.abspath("bar/"),
|
||||
os.path.abspath("bogus/"),
|
||||
],
|
||||
'quiet': 1,
|
||||
"quiet": 1,
|
||||
}
|
||||
|
||||
|
||||
def test_parse_local_config(optmanager, config_finder):
|
||||
"""Verify parsing of local config files."""
|
||||
optmanager.add_option('--exclude', parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True)
|
||||
optmanager.add_option('--ignore', parse_from_config=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option('--quiet', parse_from_config=True,
|
||||
action='count')
|
||||
optmanager.add_option(
|
||||
"--exclude",
|
||||
parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True,
|
||||
)
|
||||
optmanager.add_option(
|
||||
"--ignore", parse_from_config=True, comma_separated_list=True
|
||||
)
|
||||
optmanager.add_option("--quiet", parse_from_config=True, action="count")
|
||||
parser = config.MergedConfigParser(optmanager, config_finder)
|
||||
|
||||
with mock.patch.object(config_finder, 'local_config_files') as localcfs:
|
||||
with mock.patch.object(config_finder, "local_config_files") as localcfs:
|
||||
localcfs.return_value = [
|
||||
'tests/fixtures/config_files/cli-specified.ini'
|
||||
"tests/fixtures/config_files/cli-specified.ini"
|
||||
]
|
||||
parsed_config = parser.parse_local_config()
|
||||
|
||||
assert parsed_config == {
|
||||
'ignore': ['E123', 'W234', 'E111'],
|
||||
'exclude': [
|
||||
os.path.abspath('foo/'),
|
||||
os.path.abspath('bar/'),
|
||||
os.path.abspath('bogus/'),
|
||||
"ignore": ["E123", "W234", "E111"],
|
||||
"exclude": [
|
||||
os.path.abspath("foo/"),
|
||||
os.path.abspath("bar/"),
|
||||
os.path.abspath("bogus/"),
|
||||
],
|
||||
'quiet': 1,
|
||||
"quiet": 1,
|
||||
}
|
||||
|
||||
|
||||
def test_merge_user_and_local_config(optmanager, config_finder):
|
||||
"""Verify merging of parsed user and local config files."""
|
||||
optmanager.add_option('--exclude', parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True)
|
||||
optmanager.add_option('--ignore', parse_from_config=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option('--select', parse_from_config=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option(
|
||||
"--exclude",
|
||||
parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True,
|
||||
)
|
||||
optmanager.add_option(
|
||||
"--ignore", parse_from_config=True, comma_separated_list=True
|
||||
)
|
||||
optmanager.add_option(
|
||||
"--select", parse_from_config=True, comma_separated_list=True
|
||||
)
|
||||
parser = config.MergedConfigParser(optmanager, config_finder)
|
||||
|
||||
with mock.patch.object(config_finder, 'local_config_files') as localcfs:
|
||||
with mock.patch.object(config_finder, "local_config_files") as localcfs:
|
||||
localcfs.return_value = [
|
||||
'tests/fixtures/config_files/local-config.ini'
|
||||
"tests/fixtures/config_files/local-config.ini"
|
||||
]
|
||||
config_finder.user_config_file = ('tests/fixtures/config_files/'
|
||||
'user-config.ini')
|
||||
config_finder.user_config_file = (
|
||||
"tests/fixtures/config_files/" "user-config.ini"
|
||||
)
|
||||
parsed_config = parser.merge_user_and_local_config()
|
||||
|
||||
assert parsed_config == {
|
||||
'exclude': [
|
||||
os.path.abspath('docs/')
|
||||
],
|
||||
'ignore': ['D203'],
|
||||
'select': ['E', 'W', 'F'],
|
||||
"exclude": [os.path.abspath("docs/")],
|
||||
"ignore": ["D203"],
|
||||
"select": ["E", "W", "F"],
|
||||
}
|
||||
|
||||
|
||||
|
|
@ -154,7 +172,7 @@ def test_parse_isolates_config(optmanager):
|
|||
|
||||
def test_parse_uses_cli_config(optmanager):
|
||||
"""Verify behaviour of the parse method with a specified config."""
|
||||
config_file_value = 'foo.ini'
|
||||
config_file_value = "foo.ini"
|
||||
config_finder = mock.MagicMock()
|
||||
config_finder.config_file = config_file_value
|
||||
config_finder.ignore_config_files = False
|
||||
|
|
@ -164,62 +182,74 @@ def test_parse_uses_cli_config(optmanager):
|
|||
config_finder.cli_config.assert_called_once_with(config_file_value)
|
||||
|
||||
|
||||
@pytest.mark.parametrize('config_fixture_path', [
|
||||
'tests/fixtures/config_files/cli-specified.ini',
|
||||
'tests/fixtures/config_files/cli-specified-with-inline-comments.ini',
|
||||
'tests/fixtures/config_files/cli-specified-without-inline-comments.ini',
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"config_fixture_path",
|
||||
[
|
||||
"tests/fixtures/config_files/cli-specified.ini",
|
||||
"tests/fixtures/config_files/cli-specified-with-inline-comments.ini",
|
||||
"tests/fixtures/config_files/cli-specified-without-inline-comments.ini", # noqa: E501
|
||||
],
|
||||
)
|
||||
def test_parsed_configs_are_equivalent(
|
||||
optmanager, config_finder, config_fixture_path):
|
||||
optmanager, config_finder, config_fixture_path
|
||||
):
|
||||
"""Verify the each file matches the expected parsed output.
|
||||
|
||||
This is used to ensure our documented behaviour does not regress.
|
||||
"""
|
||||
optmanager.add_option('--exclude', parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True)
|
||||
optmanager.add_option('--ignore', parse_from_config=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option(
|
||||
"--exclude",
|
||||
parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
normalize_paths=True,
|
||||
)
|
||||
optmanager.add_option(
|
||||
"--ignore", parse_from_config=True, comma_separated_list=True
|
||||
)
|
||||
parser = config.MergedConfigParser(optmanager, config_finder)
|
||||
|
||||
with mock.patch.object(config_finder, 'local_config_files') as localcfs:
|
||||
with mock.patch.object(config_finder, "local_config_files") as localcfs:
|
||||
localcfs.return_value = [config_fixture_path]
|
||||
with mock.patch.object(config_finder,
|
||||
'user_config_file') as usercf:
|
||||
usercf.return_value = ''
|
||||
with mock.patch.object(config_finder, "user_config_file") as usercf:
|
||||
usercf.return_value = ""
|
||||
parsed_config = parser.merge_user_and_local_config()
|
||||
|
||||
assert parsed_config['ignore'] == ['E123', 'W234', 'E111']
|
||||
assert parsed_config['exclude'] == [
|
||||
os.path.abspath('foo/'),
|
||||
os.path.abspath('bar/'),
|
||||
os.path.abspath('bogus/'),
|
||||
assert parsed_config["ignore"] == ["E123", "W234", "E111"]
|
||||
assert parsed_config["exclude"] == [
|
||||
os.path.abspath("foo/"),
|
||||
os.path.abspath("bar/"),
|
||||
os.path.abspath("bogus/"),
|
||||
]
|
||||
|
||||
|
||||
@pytest.mark.parametrize('config_file', [
|
||||
'tests/fixtures/config_files/config-with-hyphenated-options.ini'
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"config_file",
|
||||
["tests/fixtures/config_files/config-with-hyphenated-options.ini"],
|
||||
)
|
||||
def test_parsed_hyphenated_and_underscored_names(
|
||||
optmanager, config_finder, config_file):
|
||||
optmanager, config_finder, config_file
|
||||
):
|
||||
"""Verify we find hyphenated option names as well as underscored.
|
||||
|
||||
This tests for options like --max-line-length and --enable-extensions
|
||||
which are able to be specified either as max-line-length or
|
||||
max_line_length in our config files.
|
||||
"""
|
||||
optmanager.add_option('--max-line-length', parse_from_config=True,
|
||||
type=int)
|
||||
optmanager.add_option('--enable-extensions', parse_from_config=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option(
|
||||
"--max-line-length", parse_from_config=True, type=int
|
||||
)
|
||||
optmanager.add_option(
|
||||
"--enable-extensions",
|
||||
parse_from_config=True,
|
||||
comma_separated_list=True,
|
||||
)
|
||||
parser = config.MergedConfigParser(optmanager, config_finder)
|
||||
|
||||
with mock.patch.object(config_finder, 'local_config_files') as localcfs:
|
||||
with mock.patch.object(config_finder, "local_config_files") as localcfs:
|
||||
localcfs.return_value = [config_file]
|
||||
with mock.patch.object(config_finder,
|
||||
'user_config_file') as usercf:
|
||||
usercf.return_value = ''
|
||||
with mock.patch.object(config_finder, "user_config_file") as usercf:
|
||||
usercf.return_value = ""
|
||||
parsed_config = parser.merge_user_and_local_config()
|
||||
|
||||
assert parsed_config['max_line_length'] == 110
|
||||
assert parsed_config['enable_extensions'] == ['H101', 'H235']
|
||||
assert parsed_config["max_line_length"] == 110
|
||||
assert parsed_config["enable_extensions"] == ["H101", "H235"]
|
||||
|
|
|
|||
|
|
@ -7,15 +7,15 @@ from flake8.formatting import default
|
|||
|
||||
def options(**kwargs):
|
||||
"""Create an argparse.Namespace instance."""
|
||||
kwargs.setdefault('output_file', None)
|
||||
kwargs.setdefault('tee', False)
|
||||
kwargs.setdefault("output_file", None)
|
||||
kwargs.setdefault("tee", False)
|
||||
return argparse.Namespace(**kwargs)
|
||||
|
||||
|
||||
def test_format_returns_nothing():
|
||||
"""Verify Nothing.format returns None."""
|
||||
formatter = default.Nothing(options())
|
||||
error = style_guide.Violation('code', 'file.py', 1, 1, 'text', '1')
|
||||
error = style_guide.Violation("code", "file.py", 1, 1, "text", "1")
|
||||
|
||||
assert formatter.format(error) is None
|
||||
|
||||
|
|
@ -23,6 +23,6 @@ def test_format_returns_nothing():
|
|||
def test_show_source_returns_nothing():
|
||||
"""Verify Nothing.show_source returns None."""
|
||||
formatter = default.Nothing(options())
|
||||
error = style_guide.Violation('code', 'file.py', 1, 1, 'text', '1')
|
||||
error = style_guide.Violation("code", "file.py", 1, 1, "text", "1")
|
||||
|
||||
assert formatter.show_source(error) is None
|
||||
|
|
|
|||
|
|
@ -10,9 +10,9 @@ from flake8.options import manager
|
|||
def test_to_argparse():
|
||||
"""Test conversion to an argparse arguments."""
|
||||
opt = manager.Option(
|
||||
short_option_name='-t',
|
||||
long_option_name='--test',
|
||||
action='count',
|
||||
short_option_name="-t",
|
||||
long_option_name="--test",
|
||||
action="count",
|
||||
parse_from_config=True,
|
||||
normalize_paths=True,
|
||||
)
|
||||
|
|
@ -20,42 +20,44 @@ def test_to_argparse():
|
|||
assert opt.parse_from_config is True
|
||||
|
||||
args, kwargs = opt.to_argparse()
|
||||
assert args == ['-t', '--test']
|
||||
assert kwargs == {'action': 'count', 'type': mock.ANY}
|
||||
assert isinstance(kwargs['type'], functools.partial)
|
||||
assert args == ["-t", "--test"]
|
||||
assert kwargs == {"action": "count", "type": mock.ANY}
|
||||
assert isinstance(kwargs["type"], functools.partial)
|
||||
|
||||
|
||||
def test_to_optparse():
|
||||
"""Test that .to_optparse() produces a useful error message."""
|
||||
with pytest.raises(AttributeError) as excinfo:
|
||||
manager.Option('--foo').to_optparse
|
||||
msg, = excinfo.value.args
|
||||
assert msg == 'to_optparse: flake8 now uses argparse'
|
||||
manager.Option("--foo").to_optparse
|
||||
(msg,) = excinfo.value.args
|
||||
assert msg == "to_optparse: flake8 now uses argparse"
|
||||
|
||||
|
||||
def test_to_argparse_creates_an_option_as_we_expect():
|
||||
"""Show that we pass all keyword args to argparse."""
|
||||
opt = manager.Option('-t', '--test', action='count')
|
||||
opt = manager.Option("-t", "--test", action="count")
|
||||
args, kwargs = opt.to_argparse()
|
||||
assert args == ['-t', '--test']
|
||||
assert kwargs == {'action': 'count'}
|
||||
assert args == ["-t", "--test"]
|
||||
assert kwargs == {"action": "count"}
|
||||
|
||||
|
||||
def test_config_name_generation():
|
||||
"""Show that we generate the config name deterministically."""
|
||||
opt = manager.Option(long_option_name='--some-very-long-option-name',
|
||||
parse_from_config=True)
|
||||
opt = manager.Option(
|
||||
long_option_name="--some-very-long-option-name",
|
||||
parse_from_config=True,
|
||||
)
|
||||
|
||||
assert opt.config_name == 'some_very_long_option_name'
|
||||
assert opt.config_name == "some_very_long_option_name"
|
||||
|
||||
|
||||
def test_config_name_needs_long_option_name():
|
||||
"""Show that we error out if the Option should be parsed from config."""
|
||||
with pytest.raises(ValueError):
|
||||
manager.Option('-s', parse_from_config=True)
|
||||
manager.Option("-s", parse_from_config=True)
|
||||
|
||||
|
||||
def test_dest_is_not_overridden():
|
||||
"""Show that we do not override custom destinations."""
|
||||
opt = manager.Option('-s', '--short', dest='something_not_short')
|
||||
assert opt.dest == 'something_not_short'
|
||||
opt = manager.Option("-s", "--short", dest="something_not_short")
|
||||
assert opt.dest == "something_not_short"
|
||||
|
|
|
|||
|
|
@ -9,13 +9,13 @@ from flake8 import utils
|
|||
from flake8.main.options import JobsArgument
|
||||
from flake8.options import manager
|
||||
|
||||
TEST_VERSION = '3.0.0b1'
|
||||
TEST_VERSION = "3.0.0b1"
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def optmanager():
|
||||
"""Generate a simple OptionManager with default test arguments."""
|
||||
return manager.OptionManager(prog='flake8', version=TEST_VERSION)
|
||||
return manager.OptionManager(prog="flake8", version=TEST_VERSION)
|
||||
|
||||
|
||||
def test_option_manager_creates_option_parser(optmanager):
|
||||
|
|
@ -27,30 +27,29 @@ def test_option_manager_including_parent_options():
|
|||
"""Verify parent options are included in the parsed options."""
|
||||
# GIVEN
|
||||
parent_parser = argparse.ArgumentParser(add_help=False)
|
||||
parent_parser.add_argument('--parent')
|
||||
parent_parser.add_argument("--parent")
|
||||
|
||||
# WHEN
|
||||
optmanager = manager.OptionManager(
|
||||
prog='flake8',
|
||||
version=TEST_VERSION,
|
||||
parents=[parent_parser])
|
||||
option, _ = optmanager.parse_args(['--parent', 'foo'])
|
||||
prog="flake8", version=TEST_VERSION, parents=[parent_parser]
|
||||
)
|
||||
option, _ = optmanager.parse_args(["--parent", "foo"])
|
||||
|
||||
# THEN
|
||||
assert option.parent == 'foo'
|
||||
assert option.parent == "foo"
|
||||
|
||||
|
||||
def test_parse_args_forwarding_default_values(optmanager):
|
||||
"""Verify default provided values are present in the final result."""
|
||||
namespace = argparse.Namespace(foo='bar')
|
||||
namespace = argparse.Namespace(foo="bar")
|
||||
options, args = optmanager.parse_args([], namespace)
|
||||
assert options.foo == 'bar'
|
||||
assert options.foo == "bar"
|
||||
|
||||
|
||||
def test_parse_args_forwarding_type_coercion(optmanager):
|
||||
"""Verify default provided values are type converted from add_option."""
|
||||
optmanager.add_option('--foo', type=int)
|
||||
namespace = argparse.Namespace(foo='5')
|
||||
optmanager.add_option("--foo", type=int)
|
||||
namespace = argparse.Namespace(foo="5")
|
||||
options, args = optmanager.parse_args([], namespace)
|
||||
assert options.foo == 5
|
||||
|
||||
|
|
@ -60,8 +59,8 @@ def test_add_option_short_option_only(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('-s', help='Test short opt')
|
||||
assert optmanager.options[0].short_option_name == '-s'
|
||||
optmanager.add_option("-s", help="Test short opt")
|
||||
assert optmanager.options[0].short_option_name == "-s"
|
||||
|
||||
|
||||
def test_add_option_long_option_only(optmanager):
|
||||
|
|
@ -69,9 +68,9 @@ def test_add_option_long_option_only(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('--long', help='Test long opt')
|
||||
optmanager.add_option("--long", help="Test long opt")
|
||||
assert optmanager.options[0].short_option_name is manager._ARG.NO
|
||||
assert optmanager.options[0].long_option_name == '--long'
|
||||
assert optmanager.options[0].long_option_name == "--long"
|
||||
|
||||
|
||||
def test_add_short_and_long_option_names(optmanager):
|
||||
|
|
@ -79,9 +78,9 @@ def test_add_short_and_long_option_names(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('-b', '--both', help='Test both opts')
|
||||
assert optmanager.options[0].short_option_name == '-b'
|
||||
assert optmanager.options[0].long_option_name == '--both'
|
||||
optmanager.add_option("-b", "--both", help="Test both opts")
|
||||
assert optmanager.options[0].short_option_name == "-b"
|
||||
assert optmanager.options[0].long_option_name == "--both"
|
||||
|
||||
|
||||
def test_add_option_with_custom_args(optmanager):
|
||||
|
|
@ -89,11 +88,11 @@ def test_add_option_with_custom_args(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('--parse', parse_from_config=True)
|
||||
optmanager.add_option('--commas', comma_separated_list=True)
|
||||
optmanager.add_option('--files', normalize_paths=True)
|
||||
optmanager.add_option("--parse", parse_from_config=True)
|
||||
optmanager.add_option("--commas", comma_separated_list=True)
|
||||
optmanager.add_option("--files", normalize_paths=True)
|
||||
|
||||
attrs = ['parse_from_config', 'comma_separated_list', 'normalize_paths']
|
||||
attrs = ["parse_from_config", "comma_separated_list", "normalize_paths"]
|
||||
for option, attr in zip(optmanager.options, attrs):
|
||||
assert getattr(option, attr) is True
|
||||
|
||||
|
|
@ -103,10 +102,10 @@ def test_parse_args_normalize_path(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('--config', normalize_paths=True)
|
||||
optmanager.add_option("--config", normalize_paths=True)
|
||||
|
||||
options, args = optmanager.parse_args(['--config', '../config.ini'])
|
||||
assert options.config == os.path.abspath('../config.ini')
|
||||
options, args = optmanager.parse_args(["--config", "../config.ini"])
|
||||
assert options.config == os.path.abspath("../config.ini")
|
||||
|
||||
|
||||
def test_parse_args_handles_comma_separated_defaults(optmanager):
|
||||
|
|
@ -114,11 +113,12 @@ def test_parse_args_handles_comma_separated_defaults(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('--exclude', default='E123,W234',
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option(
|
||||
"--exclude", default="E123,W234", comma_separated_list=True
|
||||
)
|
||||
|
||||
options, args = optmanager.parse_args([])
|
||||
assert options.exclude == ['E123', 'W234']
|
||||
assert options.exclude == ["E123", "W234"]
|
||||
|
||||
|
||||
def test_parse_args_handles_comma_separated_lists(optmanager):
|
||||
|
|
@ -126,11 +126,12 @@ def test_parse_args_handles_comma_separated_lists(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('--exclude', default='E123,W234',
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option(
|
||||
"--exclude", default="E123,W234", comma_separated_list=True
|
||||
)
|
||||
|
||||
options, args = optmanager.parse_args(['--exclude', 'E201,W111,F280'])
|
||||
assert options.exclude == ['E201', 'W111', 'F280']
|
||||
options, args = optmanager.parse_args(["--exclude", "E201,W111,F280"])
|
||||
assert options.exclude == ["E201", "W111", "F280"]
|
||||
|
||||
|
||||
def test_parse_args_normalize_paths(optmanager):
|
||||
|
|
@ -138,57 +139,61 @@ def test_parse_args_normalize_paths(optmanager):
|
|||
assert optmanager.options == []
|
||||
assert optmanager.config_options_dict == {}
|
||||
|
||||
optmanager.add_option('--extra-config', normalize_paths=True,
|
||||
comma_separated_list=True)
|
||||
optmanager.add_option(
|
||||
"--extra-config", normalize_paths=True, comma_separated_list=True
|
||||
)
|
||||
|
||||
options, args = optmanager.parse_args([
|
||||
'--extra-config', '../config.ini,tox.ini,flake8/some-other.cfg'
|
||||
])
|
||||
options, args = optmanager.parse_args(
|
||||
["--extra-config", "../config.ini,tox.ini,flake8/some-other.cfg"]
|
||||
)
|
||||
assert options.extra_config == [
|
||||
os.path.abspath('../config.ini'),
|
||||
'tox.ini',
|
||||
os.path.abspath('flake8/some-other.cfg'),
|
||||
os.path.abspath("../config.ini"),
|
||||
"tox.ini",
|
||||
os.path.abspath("flake8/some-other.cfg"),
|
||||
]
|
||||
|
||||
|
||||
def test_generate_versions(optmanager):
|
||||
"""Verify a comma-separated string is generated of registered plugins."""
|
||||
optmanager.registered_plugins = [
|
||||
manager.PluginVersion('Testing 100', '0.0.0', False),
|
||||
manager.PluginVersion('Testing 101', '0.0.0', False),
|
||||
manager.PluginVersion('Testing 300', '0.0.0', True),
|
||||
manager.PluginVersion("Testing 100", "0.0.0", False),
|
||||
manager.PluginVersion("Testing 101", "0.0.0", False),
|
||||
manager.PluginVersion("Testing 300", "0.0.0", True),
|
||||
]
|
||||
assert (optmanager.generate_versions()
|
||||
== 'Testing 100: 0.0.0, Testing 101: 0.0.0, Testing 300: 0.0.0')
|
||||
assert (
|
||||
optmanager.generate_versions()
|
||||
== "Testing 100: 0.0.0, Testing 101: 0.0.0, Testing 300: 0.0.0"
|
||||
)
|
||||
|
||||
|
||||
def test_plugins_are_sorted_in_generate_versions(optmanager):
|
||||
"""Verify we sort before joining strings in generate_versions."""
|
||||
optmanager.registered_plugins = [
|
||||
manager.PluginVersion('pyflakes', '1.5.0', False),
|
||||
manager.PluginVersion('mccabe', '0.7.0', False),
|
||||
manager.PluginVersion('pycodestyle', '2.2.0', False),
|
||||
manager.PluginVersion('flake8-docstrings', '0.6.1', False),
|
||||
manager.PluginVersion('flake8-bugbear', '2016.12.1', False),
|
||||
manager.PluginVersion("pyflakes", "1.5.0", False),
|
||||
manager.PluginVersion("mccabe", "0.7.0", False),
|
||||
manager.PluginVersion("pycodestyle", "2.2.0", False),
|
||||
manager.PluginVersion("flake8-docstrings", "0.6.1", False),
|
||||
manager.PluginVersion("flake8-bugbear", "2016.12.1", False),
|
||||
]
|
||||
assert (optmanager.generate_versions()
|
||||
== 'flake8-bugbear: 2016.12.1, '
|
||||
'flake8-docstrings: 0.6.1, '
|
||||
'mccabe: 0.7.0, '
|
||||
'pycodestyle: 2.2.0, '
|
||||
'pyflakes: 1.5.0')
|
||||
assert (
|
||||
optmanager.generate_versions() == "flake8-bugbear: 2016.12.1, "
|
||||
"flake8-docstrings: 0.6.1, "
|
||||
"mccabe: 0.7.0, "
|
||||
"pycodestyle: 2.2.0, "
|
||||
"pyflakes: 1.5.0"
|
||||
)
|
||||
|
||||
|
||||
def test_generate_versions_with_format_string(optmanager):
|
||||
"""Verify a comma-separated string is generated of registered plugins."""
|
||||
optmanager.registered_plugins.update([
|
||||
manager.PluginVersion('Testing', '0.0.0', False),
|
||||
manager.PluginVersion('Testing', '0.0.0', False),
|
||||
manager.PluginVersion('Testing', '0.0.0', False),
|
||||
])
|
||||
assert (
|
||||
optmanager.generate_versions() == 'Testing: 0.0.0'
|
||||
optmanager.registered_plugins.update(
|
||||
[
|
||||
manager.PluginVersion("Testing", "0.0.0", False),
|
||||
manager.PluginVersion("Testing", "0.0.0", False),
|
||||
manager.PluginVersion("Testing", "0.0.0", False),
|
||||
]
|
||||
)
|
||||
assert optmanager.generate_versions() == "Testing: 0.0.0"
|
||||
|
||||
|
||||
def test_update_version_string(optmanager):
|
||||
|
|
@ -197,17 +202,20 @@ def test_update_version_string(optmanager):
|
|||
assert optmanager.version_action.version == TEST_VERSION
|
||||
|
||||
optmanager.registered_plugins = [
|
||||
manager.PluginVersion('Testing 100', '0.0.0', False),
|
||||
manager.PluginVersion('Testing 101', '0.0.0', False),
|
||||
manager.PluginVersion('Testing 300', '0.0.0', False),
|
||||
manager.PluginVersion("Testing 100", "0.0.0", False),
|
||||
manager.PluginVersion("Testing 101", "0.0.0", False),
|
||||
manager.PluginVersion("Testing 300", "0.0.0", False),
|
||||
]
|
||||
|
||||
optmanager.update_version_string()
|
||||
|
||||
assert optmanager.version == TEST_VERSION
|
||||
assert (optmanager.version_action.version == TEST_VERSION
|
||||
+ ' (Testing 100: 0.0.0, Testing 101: 0.0.0, Testing 300: 0.0.0) '
|
||||
+ utils.get_python_version())
|
||||
assert (
|
||||
optmanager.version_action.version
|
||||
== TEST_VERSION
|
||||
+ " (Testing 100: 0.0.0, Testing 101: 0.0.0, Testing 300: 0.0.0) "
|
||||
+ utils.get_python_version()
|
||||
)
|
||||
|
||||
|
||||
def test_generate_epilog(optmanager):
|
||||
|
|
@ -215,14 +223,14 @@ def test_generate_epilog(optmanager):
|
|||
assert optmanager.parser.epilog is None
|
||||
|
||||
optmanager.registered_plugins = [
|
||||
manager.PluginVersion('Testing 100', '0.0.0', False),
|
||||
manager.PluginVersion('Testing 101', '0.0.0', False),
|
||||
manager.PluginVersion('Testing 300', '0.0.0', False),
|
||||
manager.PluginVersion("Testing 100", "0.0.0", False),
|
||||
manager.PluginVersion("Testing 101", "0.0.0", False),
|
||||
manager.PluginVersion("Testing 300", "0.0.0", False),
|
||||
]
|
||||
|
||||
expected_value = (
|
||||
'Installed plugins: Testing 100: 0.0.0, Testing 101: 0.0.0, Testing'
|
||||
' 300: 0.0.0'
|
||||
"Installed plugins: Testing 100: 0.0.0, Testing 101: 0.0.0, Testing"
|
||||
" 300: 0.0.0"
|
||||
)
|
||||
|
||||
optmanager.generate_epilog()
|
||||
|
|
@ -233,14 +241,14 @@ def test_extend_default_ignore(optmanager):
|
|||
"""Verify that we update the extended default ignore list."""
|
||||
assert optmanager.extended_default_ignore == set()
|
||||
|
||||
optmanager.extend_default_ignore(['T100', 'T101', 'T102'])
|
||||
assert optmanager.extended_default_ignore == {'T100', 'T101', 'T102'}
|
||||
optmanager.extend_default_ignore(["T100", "T101", "T102"])
|
||||
assert optmanager.extended_default_ignore == {"T100", "T101", "T102"}
|
||||
|
||||
|
||||
def test_parse_known_args(optmanager):
|
||||
"""Verify we ignore unknown options."""
|
||||
with mock.patch('sys.exit') as sysexit:
|
||||
optmanager.parse_known_args(['--max-complexity', '5'])
|
||||
with mock.patch("sys.exit") as sysexit:
|
||||
optmanager.parse_known_args(["--max-complexity", "5"])
|
||||
|
||||
assert sysexit.called is False
|
||||
|
||||
|
|
@ -249,101 +257,101 @@ def test_optparse_normalize_callback_option_legacy(optmanager):
|
|||
"""Test the optparse shim for `callback=`."""
|
||||
callback_foo = mock.Mock()
|
||||
optmanager.add_option(
|
||||
'--foo',
|
||||
action='callback',
|
||||
"--foo",
|
||||
action="callback",
|
||||
callback=callback_foo,
|
||||
callback_args=(1, 2),
|
||||
callback_kwargs={'a': 'b'},
|
||||
callback_kwargs={"a": "b"},
|
||||
)
|
||||
callback_bar = mock.Mock()
|
||||
optmanager.add_option(
|
||||
'--bar',
|
||||
action='callback',
|
||||
type='string',
|
||||
"--bar",
|
||||
action="callback",
|
||||
type="string",
|
||||
callback=callback_bar,
|
||||
)
|
||||
callback_baz = mock.Mock()
|
||||
optmanager.add_option(
|
||||
'--baz',
|
||||
action='callback',
|
||||
type='string',
|
||||
"--baz",
|
||||
action="callback",
|
||||
type="string",
|
||||
nargs=2,
|
||||
callback=callback_baz,
|
||||
)
|
||||
|
||||
optmanager.parse_args(['--foo', '--bar', 'bararg', '--baz', '1', '2'])
|
||||
optmanager.parse_args(["--foo", "--bar", "bararg", "--baz", "1", "2"])
|
||||
|
||||
callback_foo.assert_called_once_with(
|
||||
mock.ANY, # the option / action instance
|
||||
'--foo',
|
||||
"--foo",
|
||||
None,
|
||||
mock.ANY, # the OptionParser / ArgumentParser
|
||||
1,
|
||||
2,
|
||||
a='b',
|
||||
a="b",
|
||||
)
|
||||
callback_bar.assert_called_once_with(
|
||||
mock.ANY, # the option / action instance
|
||||
'--bar',
|
||||
'bararg',
|
||||
"--bar",
|
||||
"bararg",
|
||||
mock.ANY, # the OptionParser / ArgumentParser
|
||||
)
|
||||
callback_baz.assert_called_once_with(
|
||||
mock.ANY, # the option / action instance
|
||||
'--baz',
|
||||
('1', '2'),
|
||||
"--baz",
|
||||
("1", "2"),
|
||||
mock.ANY, # the OptionParser / ArgumentParser
|
||||
)
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
('type_s', 'input_val', 'expected'),
|
||||
("type_s", "input_val", "expected"),
|
||||
(
|
||||
('int', '5', 5),
|
||||
('long', '6', 6),
|
||||
('string', 'foo', 'foo'),
|
||||
('float', '1.5', 1.5),
|
||||
('complex', '1+5j', 1 + 5j),
|
||||
("int", "5", 5),
|
||||
("long", "6", 6),
|
||||
("string", "foo", "foo"),
|
||||
("float", "1.5", 1.5),
|
||||
("complex", "1+5j", 1 + 5j),
|
||||
# optparse allows this but does not document it
|
||||
('str', 'foo', 'foo'),
|
||||
("str", "foo", "foo"),
|
||||
),
|
||||
)
|
||||
def test_optparse_normalize_types(optmanager, type_s, input_val, expected):
|
||||
"""Test the optparse shim for type="typename"."""
|
||||
optmanager.add_option('--foo', type=type_s)
|
||||
opts, args = optmanager.parse_args(['--foo', input_val])
|
||||
optmanager.add_option("--foo", type=type_s)
|
||||
opts, args = optmanager.parse_args(["--foo", input_val])
|
||||
assert opts.foo == expected
|
||||
|
||||
|
||||
def test_optparse_normalize_choice_type(optmanager):
|
||||
"""Test the optparse shim for type="choice"."""
|
||||
optmanager.add_option('--foo', type='choice', choices=('1', '2', '3'))
|
||||
opts, args = optmanager.parse_args(['--foo', '1'])
|
||||
assert opts.foo == '1'
|
||||
optmanager.add_option("--foo", type="choice", choices=("1", "2", "3"))
|
||||
opts, args = optmanager.parse_args(["--foo", "1"])
|
||||
assert opts.foo == "1"
|
||||
# fails to parse
|
||||
with pytest.raises(SystemExit):
|
||||
optmanager.parse_args(['--foo', '4'])
|
||||
optmanager.parse_args(["--foo", "4"])
|
||||
|
||||
|
||||
def test_optparse_normalize_help(optmanager, capsys):
|
||||
"""Test the optparse shim for %default in help text."""
|
||||
optmanager.add_option('--foo', default='bar', help='default: %default')
|
||||
optmanager.add_option("--foo", default="bar", help="default: %default")
|
||||
with pytest.raises(SystemExit):
|
||||
optmanager.parse_args(['--help'])
|
||||
optmanager.parse_args(["--help"])
|
||||
out, err = capsys.readouterr()
|
||||
output = out + err
|
||||
assert 'default: bar' in output
|
||||
assert "default: bar" in output
|
||||
|
||||
|
||||
def test_optmanager_group(optmanager, capsys):
|
||||
"""Test that group(...) causes options to be assigned to a group."""
|
||||
with optmanager.group('groupname'):
|
||||
optmanager.add_option('--foo')
|
||||
with optmanager.group("groupname"):
|
||||
optmanager.add_option("--foo")
|
||||
with pytest.raises(SystemExit):
|
||||
optmanager.parse_args(['--help'])
|
||||
optmanager.parse_args(["--help"])
|
||||
out, err = capsys.readouterr()
|
||||
output = out + err
|
||||
assert '\ngroupname:\n' in output
|
||||
assert "\ngroupname:\n" in output
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
|
|
|
|||
|
|
@ -11,8 +11,8 @@ from flake8.plugins import manager
|
|||
|
||||
def test_load_plugin_fallsback_on_old_setuptools():
|
||||
"""Verify we fallback gracefully to on old versions of setuptools."""
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
|
||||
plugin.load_plugin()
|
||||
entry_point.load.assert_called_once_with()
|
||||
|
|
@ -20,8 +20,8 @@ def test_load_plugin_fallsback_on_old_setuptools():
|
|||
|
||||
def test_load_plugin_is_idempotent():
|
||||
"""Verify we use the preferred methods on new versions of setuptools."""
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
|
||||
plugin.load_plugin()
|
||||
plugin.load_plugin()
|
||||
|
|
@ -31,9 +31,9 @@ def test_load_plugin_is_idempotent():
|
|||
|
||||
def test_load_plugin_catches_and_reraises_exceptions():
|
||||
"""Verify we raise our own FailedToLoadPlugin."""
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
entry_point.load.side_effect = ValueError('Test failure')
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
entry_point.load.side_effect = ValueError("Test failure")
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
|
||||
with pytest.raises(exceptions.FailedToLoadPlugin):
|
||||
plugin.load_plugin()
|
||||
|
|
@ -41,9 +41,9 @@ def test_load_plugin_catches_and_reraises_exceptions():
|
|||
|
||||
def test_load_noncallable_plugin():
|
||||
"""Verify that we do not load a non-callable plugin."""
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
entry_point.load.return_value = mock.NonCallableMock()
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
|
||||
with pytest.raises(exceptions.FailedToLoadPlugin):
|
||||
plugin.load_plugin()
|
||||
|
|
@ -52,8 +52,8 @@ def test_load_noncallable_plugin():
|
|||
|
||||
def test_plugin_property_loads_plugin_on_first_use():
|
||||
"""Verify that we load our plugin when we first try to use it."""
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
|
||||
assert plugin.plugin is not None
|
||||
entry_point.load.assert_called_once_with()
|
||||
|
|
@ -61,14 +61,14 @@ def test_plugin_property_loads_plugin_on_first_use():
|
|||
|
||||
def test_execute_calls_plugin_with_passed_arguments():
|
||||
"""Verify that we pass arguments directly to the plugin."""
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin_obj = mock.Mock()
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
plugin._plugin = plugin_obj
|
||||
|
||||
plugin.execute('arg1', 'arg2', kwarg1='value1', kwarg2='value2')
|
||||
plugin.execute("arg1", "arg2", kwarg1="value1", kwarg2="value2")
|
||||
plugin_obj.assert_called_once_with(
|
||||
'arg1', 'arg2', kwarg1='value1', kwarg2='value2'
|
||||
"arg1", "arg2", kwarg1="value1", kwarg2="value2"
|
||||
)
|
||||
|
||||
# Extra assertions
|
||||
|
|
@ -77,23 +77,24 @@ def test_execute_calls_plugin_with_passed_arguments():
|
|||
|
||||
def test_version_proxies_to_the_plugin():
|
||||
"""Verify that we pass arguments directly to the plugin."""
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
plugin_obj = mock.Mock(spec_set=['version'])
|
||||
plugin_obj.version = 'a.b.c'
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin_obj = mock.Mock(spec_set=["version"])
|
||||
plugin_obj.version = "a.b.c"
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
plugin._plugin = plugin_obj
|
||||
|
||||
assert plugin.version == 'a.b.c'
|
||||
assert plugin.version == "a.b.c"
|
||||
|
||||
|
||||
def test_register_options():
|
||||
"""Verify we call add_options on the plugin only if it exists."""
|
||||
# Set up our mocks and Plugin object
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
plugin_obj = mock.Mock(spec_set=['name', 'version', 'add_options',
|
||||
'parse_options'])
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin_obj = mock.Mock(
|
||||
spec_set=["name", "version", "add_options", "parse_options"]
|
||||
)
|
||||
option_manager = mock.MagicMock(spec=options_manager.OptionManager)
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
plugin._plugin = plugin_obj
|
||||
|
||||
# Call the method we're testing.
|
||||
|
|
@ -106,10 +107,10 @@ def test_register_options():
|
|||
def test_register_options_checks_plugin_for_method():
|
||||
"""Verify we call add_options on the plugin only if it exists."""
|
||||
# Set up our mocks and Plugin object
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
plugin_obj = mock.Mock(spec_set=['name', 'version', 'parse_options'])
|
||||
option_manager = mock.Mock(spec=['register_plugin'])
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin_obj = mock.Mock(spec_set=["name", "version", "parse_options"])
|
||||
option_manager = mock.Mock(spec=["register_plugin"])
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
plugin._plugin = plugin_obj
|
||||
|
||||
# Call the method we're testing.
|
||||
|
|
@ -122,12 +123,13 @@ def test_register_options_checks_plugin_for_method():
|
|||
def test_provide_options():
|
||||
"""Verify we call add_options on the plugin only if it exists."""
|
||||
# Set up our mocks and Plugin object
|
||||
entry_point = mock.Mock(spec=['load'])
|
||||
plugin_obj = mock.Mock(spec_set=['name', 'version', 'add_options',
|
||||
'parse_options'])
|
||||
entry_point = mock.Mock(spec=["load"])
|
||||
plugin_obj = mock.Mock(
|
||||
spec_set=["name", "version", "add_options", "parse_options"]
|
||||
)
|
||||
option_values = argparse.Namespace(enable_extensions=[])
|
||||
option_manager = mock.Mock()
|
||||
plugin = manager.Plugin('T000', entry_point)
|
||||
plugin = manager.Plugin("T000", entry_point)
|
||||
plugin._plugin = plugin_obj
|
||||
|
||||
# Call the method we're testing.
|
||||
|
|
@ -139,10 +141,13 @@ def test_provide_options():
|
|||
)
|
||||
|
||||
|
||||
@pytest.mark.parametrize('ignore_list, code, expected_list', [
|
||||
(['E', 'W', 'F', 'C9'], 'W', ['E', 'F', 'C9']),
|
||||
(['E', 'W', 'F'], 'C9', ['E', 'W', 'F']),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"ignore_list, code, expected_list",
|
||||
[
|
||||
(["E", "W", "F", "C9"], "W", ["E", "F", "C9"]),
|
||||
(["E", "W", "F"], "C9", ["E", "W", "F"]),
|
||||
],
|
||||
)
|
||||
def test_enable(ignore_list, code, expected_list):
|
||||
"""Verify that enabling a plugin removes it from the ignore list."""
|
||||
options = mock.Mock(ignore=ignore_list)
|
||||
|
|
@ -157,8 +162,8 @@ def test_enable(ignore_list, code, expected_list):
|
|||
def test_enable_without_providing_parsed_options():
|
||||
"""Verify that enabling a plugin removes it from the ignore list."""
|
||||
optmanager = mock.Mock()
|
||||
plugin = manager.Plugin('U4', mock.Mock())
|
||||
plugin = manager.Plugin("U4", mock.Mock())
|
||||
|
||||
plugin.enable(optmanager)
|
||||
|
||||
optmanager.remove_from_default_ignore.assert_called_once_with(['U4'])
|
||||
optmanager.remove_from_default_ignore.assert_called_once_with(["U4"])
|
||||
|
|
|
|||
|
|
@ -5,54 +5,53 @@ from flake8._compat import importlib_metadata
|
|||
from flake8.plugins import manager
|
||||
|
||||
|
||||
@mock.patch.object(importlib_metadata, 'entry_points')
|
||||
@mock.patch.object(importlib_metadata, "entry_points")
|
||||
def test_calls_entrypoints_on_instantiation(entry_points_mck):
|
||||
"""Verify that we call entry_points() when we create a manager."""
|
||||
entry_points_mck.return_value = {}
|
||||
manager.PluginManager(namespace='testing.entrypoints')
|
||||
manager.PluginManager(namespace="testing.entrypoints")
|
||||
entry_points_mck.assert_called_once_with()
|
||||
|
||||
|
||||
@mock.patch.object(importlib_metadata, 'entry_points')
|
||||
@mock.patch.object(importlib_metadata, "entry_points")
|
||||
def test_calls_entrypoints_creates_plugins_automaticaly(entry_points_mck):
|
||||
"""Verify that we create Plugins on instantiation."""
|
||||
entry_points_mck.return_value = {
|
||||
'testing.entrypoints': [
|
||||
importlib_metadata.EntryPoint('T100', '', 'testing.entrypoints'),
|
||||
importlib_metadata.EntryPoint('T200', '', 'testing.entrypoints'),
|
||||
"testing.entrypoints": [
|
||||
importlib_metadata.EntryPoint("T100", "", "testing.entrypoints"),
|
||||
importlib_metadata.EntryPoint("T200", "", "testing.entrypoints"),
|
||||
],
|
||||
}
|
||||
plugin_mgr = manager.PluginManager(namespace='testing.entrypoints')
|
||||
plugin_mgr = manager.PluginManager(namespace="testing.entrypoints")
|
||||
|
||||
entry_points_mck.assert_called_once_with()
|
||||
assert 'T100' in plugin_mgr.plugins
|
||||
assert 'T200' in plugin_mgr.plugins
|
||||
assert isinstance(plugin_mgr.plugins['T100'], manager.Plugin)
|
||||
assert isinstance(plugin_mgr.plugins['T200'], manager.Plugin)
|
||||
assert "T100" in plugin_mgr.plugins
|
||||
assert "T200" in plugin_mgr.plugins
|
||||
assert isinstance(plugin_mgr.plugins["T100"], manager.Plugin)
|
||||
assert isinstance(plugin_mgr.plugins["T200"], manager.Plugin)
|
||||
|
||||
|
||||
@mock.patch.object(importlib_metadata, 'entry_points')
|
||||
@mock.patch.object(importlib_metadata, "entry_points")
|
||||
def test_handles_mapping_functions_across_plugins(entry_points_mck):
|
||||
"""Verify we can use the PluginManager call functions on all plugins."""
|
||||
entry_points_mck.return_value = {
|
||||
'testing.entrypoints': [
|
||||
importlib_metadata.EntryPoint('T100', '', 'testing.entrypoints'),
|
||||
importlib_metadata.EntryPoint('T200', '', 'testing.entrypoints'),
|
||||
"testing.entrypoints": [
|
||||
importlib_metadata.EntryPoint("T100", "", "testing.entrypoints"),
|
||||
importlib_metadata.EntryPoint("T200", "", "testing.entrypoints"),
|
||||
],
|
||||
}
|
||||
plugin_mgr = manager.PluginManager(namespace='testing.entrypoints')
|
||||
plugin_mgr = manager.PluginManager(namespace="testing.entrypoints")
|
||||
plugins = [plugin_mgr.plugins[name] for name in plugin_mgr.names]
|
||||
|
||||
assert list(plugin_mgr.map(lambda x: x)) == plugins
|
||||
|
||||
|
||||
@mock.patch.object(importlib_metadata, 'entry_points')
|
||||
@mock.patch.object(importlib_metadata, "entry_points")
|
||||
def test_local_plugins(entry_points_mck):
|
||||
"""Verify PluginManager can load given local plugins."""
|
||||
entry_points_mck.return_value = {}
|
||||
plugin_mgr = manager.PluginManager(
|
||||
namespace='testing.entrypoints',
|
||||
local_plugins=['X = path.to:Plugin']
|
||||
namespace="testing.entrypoints", local_plugins=["X = path.to:Plugin"]
|
||||
)
|
||||
|
||||
assert plugin_mgr.plugins['X'].entry_point.value == 'path.to:Plugin'
|
||||
assert plugin_mgr.plugins["X"].entry_point.value == "path.to:Plugin"
|
||||
|
|
|
|||
|
|
@ -14,8 +14,8 @@ def create_plugin_mock(raise_exception=False):
|
|||
plugin = mock.create_autospec(manager.Plugin, instance=True)
|
||||
if raise_exception:
|
||||
plugin.load_plugin.side_effect = exceptions.FailedToLoadPlugin(
|
||||
plugin_name='T101',
|
||||
exception=ValueError('Test failure'),
|
||||
plugin_name="T101",
|
||||
exception=ValueError("Test failure"),
|
||||
)
|
||||
return plugin
|
||||
|
||||
|
|
@ -28,7 +28,7 @@ def create_mapping_manager_mock(plugins):
|
|||
yield func(plugin)
|
||||
|
||||
# Mock out the PluginManager instance
|
||||
manager_mock = mock.Mock(spec=['map'])
|
||||
manager_mock = mock.Mock(spec=["map"])
|
||||
# Replace the map method
|
||||
manager_mock.map = fake_map
|
||||
return manager_mock
|
||||
|
|
@ -47,7 +47,7 @@ class FakeTestType(manager.PluginTypeManager):
|
|||
namespace = TEST_NAMESPACE
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_instantiates_a_manager(PluginManager): # noqa: N803
|
||||
"""Verify we create a PluginManager on instantiation."""
|
||||
FakeTestType()
|
||||
|
|
@ -55,26 +55,22 @@ def test_instantiates_a_manager(PluginManager): # noqa: N803
|
|||
PluginManager.assert_called_once_with(TEST_NAMESPACE, local_plugins=None)
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_proxies_names_to_manager(PluginManager): # noqa: N803
|
||||
"""Verify we proxy the names attribute."""
|
||||
PluginManager.return_value = mock.Mock(names=[
|
||||
'T100', 'T200', 'T300'
|
||||
])
|
||||
PluginManager.return_value = mock.Mock(names=["T100", "T200", "T300"])
|
||||
type_mgr = FakeTestType()
|
||||
|
||||
assert type_mgr.names == ['T100', 'T200', 'T300']
|
||||
assert type_mgr.names == ["T100", "T200", "T300"]
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_proxies_plugins_to_manager(PluginManager): # noqa: N803
|
||||
"""Verify we proxy the plugins attribute."""
|
||||
PluginManager.return_value = mock.Mock(plugins=[
|
||||
'T100', 'T200', 'T300'
|
||||
])
|
||||
PluginManager.return_value = mock.Mock(plugins=["T100", "T200", "T300"])
|
||||
type_mgr = FakeTestType()
|
||||
|
||||
assert type_mgr.plugins == ['T100', 'T200', 'T300']
|
||||
assert type_mgr.plugins == ["T100", "T200", "T300"]
|
||||
|
||||
|
||||
def test_generate_call_function():
|
||||
|
|
@ -82,21 +78,28 @@ def test_generate_call_function():
|
|||
optmanager = object()
|
||||
plugin = mock.Mock(method_name=lambda x: x)
|
||||
func = manager.PluginTypeManager._generate_call_function(
|
||||
'method_name', optmanager,
|
||||
"method_name",
|
||||
optmanager,
|
||||
)
|
||||
|
||||
assert callable(func)
|
||||
assert func(plugin) is optmanager
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_load_plugins(PluginManager): # noqa: N803
|
||||
"""Verify load plugins loads *every* plugin."""
|
||||
# Create a bunch of fake plugins
|
||||
plugins = [create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock()]
|
||||
plugins = [
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
]
|
||||
# Return our PluginManager mock
|
||||
PluginManager.return_value = create_mapping_manager_mock(plugins)
|
||||
|
||||
|
|
@ -109,13 +112,19 @@ def test_load_plugins(PluginManager): # noqa: N803
|
|||
assert type_mgr.plugins_loaded is True
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_load_plugins_fails(PluginManager): # noqa: N803
|
||||
"""Verify load plugins bubbles up exceptions."""
|
||||
plugins = [create_plugin_mock(), create_plugin_mock(True),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock()]
|
||||
plugins = [
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(True),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
]
|
||||
# Return our PluginManager mock
|
||||
PluginManager.return_value = create_mapping_manager_mock(plugins)
|
||||
|
||||
|
|
@ -133,13 +142,19 @@ def test_load_plugins_fails(PluginManager): # noqa: N803
|
|||
assert plugin.load_plugin.called is False
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_register_options(PluginManager): # noqa: N803
|
||||
"""Test that we map over every plugin to register options."""
|
||||
plugins = [create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock()]
|
||||
plugins = [
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
]
|
||||
# Return our PluginManager mock
|
||||
PluginManager.return_value = create_mapping_manager_mock(plugins)
|
||||
optmanager = object()
|
||||
|
|
@ -151,13 +166,19 @@ def test_register_options(PluginManager): # noqa: N803
|
|||
plugin.register_options.assert_called_with(optmanager)
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_provide_options(PluginManager): # noqa: N803
|
||||
"""Test that we map over every plugin to provide parsed options."""
|
||||
plugins = [create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock(),
|
||||
create_plugin_mock(), create_plugin_mock()]
|
||||
plugins = [
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
create_plugin_mock(),
|
||||
]
|
||||
# Return our PluginManager mock
|
||||
PluginManager.return_value = create_mapping_manager_mock(plugins)
|
||||
optmanager = object()
|
||||
|
|
@ -167,32 +188,30 @@ def test_provide_options(PluginManager): # noqa: N803
|
|||
type_mgr.provide_options(optmanager, options, [])
|
||||
|
||||
for plugin in plugins:
|
||||
plugin.provide_options.assert_called_with(optmanager,
|
||||
options,
|
||||
[])
|
||||
plugin.provide_options.assert_called_with(optmanager, options, [])
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_proxy_contains_to_managers_plugins_dict(PluginManager): # noqa: N803
|
||||
"""Verify that we proxy __contains__ to the manager's dictionary."""
|
||||
plugins = {'T10%i' % i: create_plugin_mock() for i in range(8)}
|
||||
plugins = {"T10%i" % i: create_plugin_mock() for i in range(8)}
|
||||
# Return our PluginManager mock
|
||||
PluginManager.return_value = create_manager_with_plugins(plugins)
|
||||
|
||||
type_mgr = FakeTestType()
|
||||
for i in range(8):
|
||||
key = 'T10%i' % i
|
||||
key = "T10%i" % i
|
||||
assert key in type_mgr
|
||||
|
||||
|
||||
@mock.patch('flake8.plugins.manager.PluginManager')
|
||||
@mock.patch("flake8.plugins.manager.PluginManager")
|
||||
def test_proxies_getitem_to_managers_plugins_dict(PluginManager): # noqa: N803
|
||||
"""Verify that we can use the PluginTypeManager like a dictionary."""
|
||||
plugins = {'T10%i' % i: create_plugin_mock() for i in range(8)}
|
||||
plugins = {"T10%i" % i: create_plugin_mock() for i in range(8)}
|
||||
# Return our PluginManager mock
|
||||
PluginManager.return_value = create_manager_with_plugins(plugins)
|
||||
|
||||
type_mgr = FakeTestType()
|
||||
for i in range(8):
|
||||
key = 'T10%i' % i
|
||||
key = "T10%i" % i
|
||||
assert type_mgr[key] is plugins[key]
|
||||
|
|
|
|||
|
|
@ -18,14 +18,14 @@ def test_all_pyflakes_messages_have_flake8_codes_assigned():
|
|||
|
||||
def test_undefined_local_code():
|
||||
"""In pyflakes 2.1.0 this code's string formatting was changed."""
|
||||
src = '''\
|
||||
src = """\
|
||||
import sys
|
||||
|
||||
def f():
|
||||
sys = sys
|
||||
'''
|
||||
"""
|
||||
tree = ast.parse(src)
|
||||
checker = pyflakes_shim.FlakesChecker(tree, (), 't.py')
|
||||
checker = pyflakes_shim.FlakesChecker(tree, (), "t.py")
|
||||
message_texts = [s for _, _, s, _ in checker.run()]
|
||||
assert message_texts == [
|
||||
"F823 local variable 'sys' defined in enclosing scope on line 1 referenced before assignment", # noqa: E501
|
||||
|
|
|
|||
|
|
@ -4,19 +4,19 @@ import pytest
|
|||
from flake8 import statistics as stats
|
||||
from flake8 import style_guide
|
||||
|
||||
DEFAULT_ERROR_CODE = 'E100'
|
||||
DEFAULT_FILENAME = 'file.py'
|
||||
DEFAULT_TEXT = 'Default text'
|
||||
DEFAULT_ERROR_CODE = "E100"
|
||||
DEFAULT_FILENAME = "file.py"
|
||||
DEFAULT_TEXT = "Default text"
|
||||
|
||||
|
||||
def make_error(**kwargs):
|
||||
"""Create errors with a bunch of default values."""
|
||||
return style_guide.Violation(
|
||||
code=kwargs.pop('code', DEFAULT_ERROR_CODE),
|
||||
filename=kwargs.pop('filename', DEFAULT_FILENAME),
|
||||
line_number=kwargs.pop('line_number', 1),
|
||||
column_number=kwargs.pop('column_number', 1),
|
||||
text=kwargs.pop('text', DEFAULT_TEXT),
|
||||
code=kwargs.pop("code", DEFAULT_ERROR_CODE),
|
||||
filename=kwargs.pop("filename", DEFAULT_FILENAME),
|
||||
line_number=kwargs.pop("line_number", 1),
|
||||
column_number=kwargs.pop("column_number", 1),
|
||||
text=kwargs.pop("text", DEFAULT_TEXT),
|
||||
physical_line=None,
|
||||
)
|
||||
|
||||
|
|
@ -29,26 +29,29 @@ def test_key_creation():
|
|||
assert key.code == DEFAULT_ERROR_CODE
|
||||
|
||||
|
||||
@pytest.mark.parametrize('code, filename, args, expected_result', [
|
||||
# Error prefix matches
|
||||
('E123', 'file000.py', ('E', None), True),
|
||||
('E123', 'file000.py', ('E1', None), True),
|
||||
('E123', 'file000.py', ('E12', None), True),
|
||||
('E123', 'file000.py', ('E123', None), True),
|
||||
# Error prefix and filename match
|
||||
('E123', 'file000.py', ('E', 'file000.py'), True),
|
||||
('E123', 'file000.py', ('E1', 'file000.py'), True),
|
||||
('E123', 'file000.py', ('E12', 'file000.py'), True),
|
||||
('E123', 'file000.py', ('E123', 'file000.py'), True),
|
||||
# Error prefix does not match
|
||||
('E123', 'file000.py', ('W', None), False),
|
||||
# Error prefix matches but filename does not
|
||||
('E123', 'file000.py', ('E', 'file001.py'), False),
|
||||
# Error prefix does not match but filename does
|
||||
('E123', 'file000.py', ('W', 'file000.py'), False),
|
||||
# Neither error prefix match nor filename
|
||||
('E123', 'file000.py', ('W', 'file001.py'), False),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"code, filename, args, expected_result",
|
||||
[
|
||||
# Error prefix matches
|
||||
("E123", "file000.py", ("E", None), True),
|
||||
("E123", "file000.py", ("E1", None), True),
|
||||
("E123", "file000.py", ("E12", None), True),
|
||||
("E123", "file000.py", ("E123", None), True),
|
||||
# Error prefix and filename match
|
||||
("E123", "file000.py", ("E", "file000.py"), True),
|
||||
("E123", "file000.py", ("E1", "file000.py"), True),
|
||||
("E123", "file000.py", ("E12", "file000.py"), True),
|
||||
("E123", "file000.py", ("E123", "file000.py"), True),
|
||||
# Error prefix does not match
|
||||
("E123", "file000.py", ("W", None), False),
|
||||
# Error prefix matches but filename does not
|
||||
("E123", "file000.py", ("E", "file001.py"), False),
|
||||
# Error prefix does not match but filename does
|
||||
("E123", "file000.py", ("W", "file000.py"), False),
|
||||
# Neither error prefix match nor filename
|
||||
("E123", "file000.py", ("W", "file001.py"), False),
|
||||
],
|
||||
)
|
||||
def test_key_matching(code, filename, args, expected_result):
|
||||
"""Verify Key#matches behaves as we expect with fthe above input."""
|
||||
key = stats.Key.create_from(make_error(code=code, filename=filename))
|
||||
|
|
@ -75,7 +78,7 @@ def test_statistic_increment():
|
|||
def test_recording_statistics():
|
||||
"""Verify that we appropriately create a new Statistic and store it."""
|
||||
aggregator = stats.Statistics()
|
||||
assert list(aggregator.statistics_for('E')) == []
|
||||
assert list(aggregator.statistics_for("E")) == []
|
||||
aggregator.record(make_error())
|
||||
storage = aggregator._store
|
||||
for key, value in storage.items():
|
||||
|
|
@ -88,9 +91,9 @@ def test_recording_statistics():
|
|||
def test_statistics_for_single_record():
|
||||
"""Show we can retrieve the only statistic recorded."""
|
||||
aggregator = stats.Statistics()
|
||||
assert list(aggregator.statistics_for('E')) == []
|
||||
assert list(aggregator.statistics_for("E")) == []
|
||||
aggregator.record(make_error())
|
||||
statistics = list(aggregator.statistics_for('E'))
|
||||
statistics = list(aggregator.statistics_for("E"))
|
||||
assert len(statistics) == 1
|
||||
assert isinstance(statistics[0], stats.Statistic)
|
||||
|
||||
|
|
@ -98,11 +101,11 @@ def test_statistics_for_single_record():
|
|||
def test_statistics_for_filters_by_filename():
|
||||
"""Show we can retrieve the only statistic recorded."""
|
||||
aggregator = stats.Statistics()
|
||||
assert list(aggregator.statistics_for('E')) == []
|
||||
assert list(aggregator.statistics_for("E")) == []
|
||||
aggregator.record(make_error())
|
||||
aggregator.record(make_error(filename='example.py'))
|
||||
aggregator.record(make_error(filename="example.py"))
|
||||
|
||||
statistics = list(aggregator.statistics_for('E', DEFAULT_FILENAME))
|
||||
statistics = list(aggregator.statistics_for("E", DEFAULT_FILENAME))
|
||||
assert len(statistics) == 1
|
||||
assert isinstance(statistics[0], stats.Statistic)
|
||||
|
||||
|
|
@ -111,11 +114,11 @@ def test_statistic_for_retrieves_more_than_one_value():
|
|||
"""Show this works for more than a couple statistic values."""
|
||||
aggregator = stats.Statistics()
|
||||
for i in range(50):
|
||||
aggregator.record(make_error(code=f'E1{i:02d}'))
|
||||
aggregator.record(make_error(code=f'W2{i:02d}'))
|
||||
aggregator.record(make_error(code=f"E1{i:02d}"))
|
||||
aggregator.record(make_error(code=f"W2{i:02d}"))
|
||||
|
||||
statistics = list(aggregator.statistics_for('E'))
|
||||
statistics = list(aggregator.statistics_for("E"))
|
||||
assert len(statistics) == 50
|
||||
|
||||
statistics = list(aggregator.statistics_for('W22'))
|
||||
statistics = list(aggregator.statistics_for("W22"))
|
||||
assert len(statistics) == 10
|
||||
|
|
|
|||
|
|
@ -14,86 +14,92 @@ from flake8.plugins import manager as plugin_manager
|
|||
RELATIVE_PATHS = ["flake8", "pep8", "pyflakes", "mccabe"]
|
||||
|
||||
|
||||
@pytest.mark.parametrize("value,expected", [
|
||||
("E123,\n\tW234,\n E206", ["E123", "W234", "E206"]),
|
||||
("E123,W234,E206", ["E123", "W234", "E206"]),
|
||||
("E123 W234 E206", ["E123", "W234", "E206"]),
|
||||
("E123\nW234 E206", ["E123", "W234", "E206"]),
|
||||
("E123\nW234\nE206", ["E123", "W234", "E206"]),
|
||||
("E123,W234,E206,", ["E123", "W234", "E206"]),
|
||||
("E123,W234,E206, ,\n", ["E123", "W234", "E206"]),
|
||||
("E123,W234,,E206,,", ["E123", "W234", "E206"]),
|
||||
("E123, W234,, E206,,", ["E123", "W234", "E206"]),
|
||||
("E123,,W234,,E206,,", ["E123", "W234", "E206"]),
|
||||
("", []),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"value,expected",
|
||||
[
|
||||
("E123,\n\tW234,\n E206", ["E123", "W234", "E206"]),
|
||||
("E123,W234,E206", ["E123", "W234", "E206"]),
|
||||
("E123 W234 E206", ["E123", "W234", "E206"]),
|
||||
("E123\nW234 E206", ["E123", "W234", "E206"]),
|
||||
("E123\nW234\nE206", ["E123", "W234", "E206"]),
|
||||
("E123,W234,E206,", ["E123", "W234", "E206"]),
|
||||
("E123,W234,E206, ,\n", ["E123", "W234", "E206"]),
|
||||
("E123,W234,,E206,,", ["E123", "W234", "E206"]),
|
||||
("E123, W234,, E206,,", ["E123", "W234", "E206"]),
|
||||
("E123,,W234,,E206,,", ["E123", "W234", "E206"]),
|
||||
("", []),
|
||||
],
|
||||
)
|
||||
def test_parse_comma_separated_list(value, expected):
|
||||
"""Verify that similar inputs produce identical outputs."""
|
||||
assert utils.parse_comma_separated_list(value) == expected
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
('value', 'expected'),
|
||||
("value", "expected"),
|
||||
(
|
||||
# empty option configures nothing
|
||||
('', []), (' ', []), ('\n\n\n', []),
|
||||
("", []),
|
||||
(" ", []),
|
||||
("\n\n\n", []),
|
||||
# basic case
|
||||
(
|
||||
'f.py:E123',
|
||||
[('f.py', ['E123'])],
|
||||
"f.py:E123",
|
||||
[("f.py", ["E123"])],
|
||||
),
|
||||
# multiple filenames, multiple codes
|
||||
(
|
||||
'f.py,g.py:E,F',
|
||||
[('f.py', ['E', 'F']), ('g.py', ['E', 'F'])],
|
||||
"f.py,g.py:E,F",
|
||||
[("f.py", ["E", "F"]), ("g.py", ["E", "F"])],
|
||||
),
|
||||
# demonstrate that whitespace is not important around tokens
|
||||
(
|
||||
' f.py , g.py : E , F ',
|
||||
[('f.py', ['E', 'F']), ('g.py', ['E', 'F'])],
|
||||
" f.py , g.py : E , F ",
|
||||
[("f.py", ["E", "F"]), ("g.py", ["E", "F"])],
|
||||
),
|
||||
# whitespace can separate groups of configuration
|
||||
(
|
||||
'f.py:E g.py:F',
|
||||
[('f.py', ['E']), ('g.py', ['F'])],
|
||||
"f.py:E g.py:F",
|
||||
[("f.py", ["E"]), ("g.py", ["F"])],
|
||||
),
|
||||
# newlines can separate groups of configuration
|
||||
(
|
||||
'f.py: E\ng.py: F\n',
|
||||
[('f.py', ['E']), ('g.py', ['F'])],
|
||||
"f.py: E\ng.py: F\n",
|
||||
[("f.py", ["E"]), ("g.py", ["F"])],
|
||||
),
|
||||
# whitespace can be used in place of commas
|
||||
(
|
||||
'f.py g.py: E F',
|
||||
[('f.py', ['E', 'F']), ('g.py', ['E', 'F'])],
|
||||
"f.py g.py: E F",
|
||||
[("f.py", ["E", "F"]), ("g.py", ["E", "F"])],
|
||||
),
|
||||
# go ahead, indent your codes
|
||||
(
|
||||
'f.py:\n E,F\ng.py:\n G,H',
|
||||
[('f.py', ['E', 'F']), ('g.py', ['G', 'H'])],
|
||||
"f.py:\n E,F\ng.py:\n G,H",
|
||||
[("f.py", ["E", "F"]), ("g.py", ["G", "H"])],
|
||||
),
|
||||
# capitalized filenames are ok too
|
||||
(
|
||||
'F.py,G.py: F,G',
|
||||
[('F.py', ['F', 'G']), ('G.py', ['F', 'G'])],
|
||||
"F.py,G.py: F,G",
|
||||
[("F.py", ["F", "G"]), ("G.py", ["F", "G"])],
|
||||
),
|
||||
# it's easier to allow zero filenames or zero codes than forbid it
|
||||
(':E', []), ('f.py:', []),
|
||||
(':E f.py:F', [('f.py', ['F'])]),
|
||||
('f.py: g.py:F', [('g.py', ['F'])]),
|
||||
('f.py:E:', []),
|
||||
('f.py:E.py:', []),
|
||||
('f.py:Eg.py:F', [('Eg.py', ['F'])]),
|
||||
(":E", []),
|
||||
("f.py:", []),
|
||||
(":E f.py:F", [("f.py", ["F"])]),
|
||||
("f.py: g.py:F", [("g.py", ["F"])]),
|
||||
("f.py:E:", []),
|
||||
("f.py:E.py:", []),
|
||||
("f.py:Eg.py:F", [("Eg.py", ["F"])]),
|
||||
# sequences are also valid (?)
|
||||
(
|
||||
['f.py:E,F', 'g.py:G,H'],
|
||||
[('f.py', ['E', 'F']), ('g.py', ['G', 'H'])],
|
||||
["f.py:E,F", "g.py:G,H"],
|
||||
[("f.py", ["E", "F"]), ("g.py", ["G", "H"])],
|
||||
),
|
||||
# six-digits codes are allowed
|
||||
(
|
||||
'f.py: ABC123',
|
||||
[('f.py', ['ABC123'])],
|
||||
)
|
||||
"f.py: ABC123",
|
||||
[("f.py", ["ABC123"])],
|
||||
),
|
||||
),
|
||||
)
|
||||
def test_parse_files_to_codes_mapping(value, expected):
|
||||
|
|
@ -102,16 +108,19 @@ def test_parse_files_to_codes_mapping(value, expected):
|
|||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'value',
|
||||
"value",
|
||||
(
|
||||
# code while looking for filenames
|
||||
'E123', 'f.py,E123', 'f.py E123',
|
||||
"E123",
|
||||
"f.py,E123",
|
||||
"f.py E123",
|
||||
# eof while looking for filenames
|
||||
'f.py', 'f.py:E,g.py'
|
||||
"f.py",
|
||||
"f.py:E,g.py"
|
||||
# colon while looking for codes
|
||||
'f.py::',
|
||||
"f.py::",
|
||||
# no separator between
|
||||
'f.py:E1F1',
|
||||
"f.py:E1F1",
|
||||
),
|
||||
)
|
||||
def test_invalid_file_list(value):
|
||||
|
|
@ -120,22 +129,32 @@ def test_invalid_file_list(value):
|
|||
utils.parse_files_to_codes_mapping(value)
|
||||
|
||||
|
||||
@pytest.mark.parametrize("value,expected", [
|
||||
("flake8", "flake8"),
|
||||
("../flake8", os.path.abspath("../flake8")),
|
||||
("flake8/", os.path.abspath("flake8")),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"value,expected",
|
||||
[
|
||||
("flake8", "flake8"),
|
||||
("../flake8", os.path.abspath("../flake8")),
|
||||
("flake8/", os.path.abspath("flake8")),
|
||||
],
|
||||
)
|
||||
def test_normalize_path(value, expected):
|
||||
"""Verify that we normalize paths provided to the tool."""
|
||||
assert utils.normalize_path(value) == expected
|
||||
|
||||
|
||||
@pytest.mark.parametrize("value,expected", [
|
||||
(["flake8", "pep8", "pyflakes", "mccabe"],
|
||||
["flake8", "pep8", "pyflakes", "mccabe"]),
|
||||
(["../flake8", "../pep8", "../pyflakes", "../mccabe"],
|
||||
[os.path.abspath(f"../{p}") for p in RELATIVE_PATHS]),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"value,expected",
|
||||
[
|
||||
(
|
||||
["flake8", "pep8", "pyflakes", "mccabe"],
|
||||
["flake8", "pep8", "pyflakes", "mccabe"],
|
||||
),
|
||||
(
|
||||
["../flake8", "../pep8", "../pyflakes", "../mccabe"],
|
||||
[os.path.abspath(f"../{p}") for p in RELATIVE_PATHS],
|
||||
),
|
||||
],
|
||||
)
|
||||
def test_normalize_paths(value, expected):
|
||||
"""Verify we normalizes a sequence of paths provided to the tool."""
|
||||
assert utils.normalize_paths(value) == expected
|
||||
|
|
@ -143,19 +162,22 @@ def test_normalize_paths(value, expected):
|
|||
|
||||
def test_is_windows_checks_for_nt():
|
||||
"""Verify that we correctly detect Windows."""
|
||||
with mock.patch.object(os, 'name', 'nt'):
|
||||
with mock.patch.object(os, "name", "nt"):
|
||||
assert utils.is_windows() is True
|
||||
|
||||
with mock.patch.object(os, 'name', 'posix'):
|
||||
with mock.patch.object(os, "name", "posix"):
|
||||
assert utils.is_windows() is False
|
||||
|
||||
|
||||
@pytest.mark.parametrize('filename,patterns,expected', [
|
||||
('foo.py', [], True),
|
||||
('foo.py', ['*.pyc'], False),
|
||||
('foo.pyc', ['*.pyc'], True),
|
||||
('foo.pyc', ['*.swp', '*.pyc', '*.py'], True),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"filename,patterns,expected",
|
||||
[
|
||||
("foo.py", [], True),
|
||||
("foo.py", ["*.pyc"], False),
|
||||
("foo.pyc", ["*.pyc"], True),
|
||||
("foo.pyc", ["*.swp", "*.pyc", "*.py"], True),
|
||||
],
|
||||
)
|
||||
def test_fnmatch(filename, patterns, expected):
|
||||
"""Verify that our fnmatch wrapper works as expected."""
|
||||
assert utils.fnmatch(filename, patterns) is expected
|
||||
|
|
@ -165,104 +187,110 @@ def test_fnmatch(filename, patterns, expected):
|
|||
def files_dir(tmpdir):
|
||||
"""Create test dir for testing filenames_from."""
|
||||
with tmpdir.as_cwd():
|
||||
tmpdir.join('a/b/c.py').ensure()
|
||||
tmpdir.join('a/b/d.py').ensure()
|
||||
tmpdir.join('a/b/e/f.py').ensure()
|
||||
tmpdir.join("a/b/c.py").ensure()
|
||||
tmpdir.join("a/b/d.py").ensure()
|
||||
tmpdir.join("a/b/e/f.py").ensure()
|
||||
yield tmpdir
|
||||
|
||||
|
||||
def _normpath(s):
|
||||
return s.replace('/', os.sep)
|
||||
return s.replace("/", os.sep)
|
||||
|
||||
|
||||
def _normpaths(pths):
|
||||
return {_normpath(pth) for pth in pths}
|
||||
|
||||
|
||||
@pytest.mark.usefixtures('files_dir')
|
||||
@pytest.mark.usefixtures("files_dir")
|
||||
def test_filenames_from_a_directory():
|
||||
"""Verify that filenames_from walks a directory."""
|
||||
filenames = set(utils.filenames_from(_normpath('a/b/')))
|
||||
filenames = set(utils.filenames_from(_normpath("a/b/")))
|
||||
# should include all files
|
||||
expected = _normpaths(('a/b/c.py', 'a/b/d.py', 'a/b/e/f.py'))
|
||||
expected = _normpaths(("a/b/c.py", "a/b/d.py", "a/b/e/f.py"))
|
||||
assert filenames == expected
|
||||
|
||||
|
||||
@pytest.mark.usefixtures('files_dir')
|
||||
@pytest.mark.usefixtures("files_dir")
|
||||
def test_filenames_from_a_directory_with_a_predicate():
|
||||
"""Verify that predicates filter filenames_from."""
|
||||
filenames = set(utils.filenames_from(
|
||||
arg=_normpath('a/b/'),
|
||||
predicate=lambda path: path.endswith(_normpath('b/c.py')),
|
||||
))
|
||||
filenames = set(
|
||||
utils.filenames_from(
|
||||
arg=_normpath("a/b/"),
|
||||
predicate=lambda path: path.endswith(_normpath("b/c.py")),
|
||||
)
|
||||
)
|
||||
# should not include c.py
|
||||
expected = _normpaths(('a/b/d.py', 'a/b/e/f.py'))
|
||||
expected = _normpaths(("a/b/d.py", "a/b/e/f.py"))
|
||||
assert filenames == expected
|
||||
|
||||
|
||||
@pytest.mark.usefixtures('files_dir')
|
||||
@pytest.mark.usefixtures("files_dir")
|
||||
def test_filenames_from_a_directory_with_a_predicate_from_the_current_dir():
|
||||
"""Verify that predicates filter filenames_from."""
|
||||
filenames = set(utils.filenames_from(
|
||||
arg=_normpath('./a/b'),
|
||||
predicate=lambda path: path == 'c.py',
|
||||
))
|
||||
filenames = set(
|
||||
utils.filenames_from(
|
||||
arg=_normpath("./a/b"),
|
||||
predicate=lambda path: path == "c.py",
|
||||
)
|
||||
)
|
||||
# none should have matched the predicate so all returned
|
||||
expected = _normpaths(('./a/b/c.py', './a/b/d.py', './a/b/e/f.py'))
|
||||
expected = _normpaths(("./a/b/c.py", "./a/b/d.py", "./a/b/e/f.py"))
|
||||
assert filenames == expected
|
||||
|
||||
|
||||
@pytest.mark.usefixtures('files_dir')
|
||||
@pytest.mark.usefixtures("files_dir")
|
||||
def test_filenames_from_a_single_file():
|
||||
"""Verify that we simply yield that filename."""
|
||||
filenames = set(utils.filenames_from(_normpath('a/b/c.py')))
|
||||
assert filenames == {_normpath('a/b/c.py')}
|
||||
filenames = set(utils.filenames_from(_normpath("a/b/c.py")))
|
||||
assert filenames == {_normpath("a/b/c.py")}
|
||||
|
||||
|
||||
def test_filenames_from_a_single_file_does_not_exist():
|
||||
"""Verify that a passed filename which does not exist is returned back."""
|
||||
filenames = set(utils.filenames_from(_normpath('d/n/e.py')))
|
||||
assert filenames == {_normpath('d/n/e.py')}
|
||||
filenames = set(utils.filenames_from(_normpath("d/n/e.py")))
|
||||
assert filenames == {_normpath("d/n/e.py")}
|
||||
|
||||
|
||||
def test_filenames_from_exclude_doesnt_exclude_directory_names(tmpdir):
|
||||
"""Verify that we don't greedily exclude subdirs."""
|
||||
tmpdir.join('1').ensure_dir().join('dont_return_me.py').ensure()
|
||||
tmpdir.join('2').join('1').ensure_dir().join('return_me.py').ensure()
|
||||
exclude = [tmpdir.join('1').strpath]
|
||||
tmpdir.join("1").ensure_dir().join("dont_return_me.py").ensure()
|
||||
tmpdir.join("2").join("1").ensure_dir().join("return_me.py").ensure()
|
||||
exclude = [tmpdir.join("1").strpath]
|
||||
|
||||
# This acts similar to src.flake8.checker.is_path_excluded
|
||||
def predicate(pth):
|
||||
return utils.fnmatch(os.path.abspath(pth), exclude)
|
||||
|
||||
with tmpdir.as_cwd():
|
||||
filenames = list(utils.filenames_from('.', predicate))
|
||||
assert filenames == [os.path.join('.', '2', '1', 'return_me.py')]
|
||||
filenames = list(utils.filenames_from(".", predicate))
|
||||
assert filenames == [os.path.join(".", "2", "1", "return_me.py")]
|
||||
|
||||
|
||||
def test_parameters_for_class_plugin():
|
||||
"""Verify that we can retrieve the parameters for a class plugin."""
|
||||
|
||||
class FakeCheck:
|
||||
def __init__(self, tree):
|
||||
raise NotImplementedError
|
||||
|
||||
plugin = plugin_manager.Plugin('plugin-name', object())
|
||||
plugin = plugin_manager.Plugin("plugin-name", object())
|
||||
plugin._plugin = FakeCheck
|
||||
assert utils.parameters_for(plugin) == {'tree': True}
|
||||
assert utils.parameters_for(plugin) == {"tree": True}
|
||||
|
||||
|
||||
def test_parameters_for_function_plugin():
|
||||
"""Verify that we retrieve the parameters for a function plugin."""
|
||||
|
||||
def fake_plugin(physical_line, self, tree, optional=None):
|
||||
raise NotImplementedError
|
||||
|
||||
plugin = plugin_manager.Plugin('plugin-name', object())
|
||||
plugin = plugin_manager.Plugin("plugin-name", object())
|
||||
plugin._plugin = fake_plugin
|
||||
assert utils.parameters_for(plugin) == {
|
||||
'physical_line': True,
|
||||
'self': True,
|
||||
'tree': True,
|
||||
'optional': False,
|
||||
"physical_line": True,
|
||||
"self": True,
|
||||
"tree": True,
|
||||
"optional": False,
|
||||
}
|
||||
|
||||
|
||||
|
|
@ -273,29 +301,32 @@ def read_diff_file(filename):
|
|||
return content
|
||||
|
||||
|
||||
SINGLE_FILE_DIFF = read_diff_file('tests/fixtures/diffs/single_file_diff')
|
||||
SINGLE_FILE_DIFF = read_diff_file("tests/fixtures/diffs/single_file_diff")
|
||||
SINGLE_FILE_INFO = {
|
||||
'flake8/utils.py': set(range(75, 83)).union(set(range(84, 94))),
|
||||
"flake8/utils.py": set(range(75, 83)).union(set(range(84, 94))),
|
||||
}
|
||||
TWO_FILE_DIFF = read_diff_file('tests/fixtures/diffs/two_file_diff')
|
||||
TWO_FILE_DIFF = read_diff_file("tests/fixtures/diffs/two_file_diff")
|
||||
TWO_FILE_INFO = {
|
||||
'flake8/utils.py': set(range(75, 83)).union(set(range(84, 94))),
|
||||
'tests/unit/test_utils.py': set(range(115, 128)),
|
||||
"flake8/utils.py": set(range(75, 83)).union(set(range(84, 94))),
|
||||
"tests/unit/test_utils.py": set(range(115, 128)),
|
||||
}
|
||||
MULTI_FILE_DIFF = read_diff_file('tests/fixtures/diffs/multi_file_diff')
|
||||
MULTI_FILE_DIFF = read_diff_file("tests/fixtures/diffs/multi_file_diff")
|
||||
MULTI_FILE_INFO = {
|
||||
'flake8/utils.py': set(range(75, 83)).union(set(range(84, 94))),
|
||||
'tests/unit/test_utils.py': set(range(115, 129)),
|
||||
'tests/fixtures/diffs/single_file_diff': set(range(1, 28)),
|
||||
'tests/fixtures/diffs/two_file_diff': set(range(1, 46)),
|
||||
"flake8/utils.py": set(range(75, 83)).union(set(range(84, 94))),
|
||||
"tests/unit/test_utils.py": set(range(115, 129)),
|
||||
"tests/fixtures/diffs/single_file_diff": set(range(1, 28)),
|
||||
"tests/fixtures/diffs/two_file_diff": set(range(1, 46)),
|
||||
}
|
||||
|
||||
|
||||
@pytest.mark.parametrize("diff, parsed_diff", [
|
||||
(SINGLE_FILE_DIFF, SINGLE_FILE_INFO),
|
||||
(TWO_FILE_DIFF, TWO_FILE_INFO),
|
||||
(MULTI_FILE_DIFF, MULTI_FILE_INFO),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"diff, parsed_diff",
|
||||
[
|
||||
(SINGLE_FILE_DIFF, SINGLE_FILE_INFO),
|
||||
(TWO_FILE_DIFF, TWO_FILE_INFO),
|
||||
(MULTI_FILE_DIFF, MULTI_FILE_INFO),
|
||||
],
|
||||
)
|
||||
def test_parse_unified_diff(diff, parsed_diff):
|
||||
"""Verify that what we parse from a diff matches expectations."""
|
||||
assert utils.parse_unified_diff(diff) == parsed_diff
|
||||
|
|
@ -304,19 +335,19 @@ def test_parse_unified_diff(diff, parsed_diff):
|
|||
def test_matches_filename_for_excluding_dotfiles():
|
||||
"""Verify that `.` and `..` are not matched by `.*`."""
|
||||
logger = logging.Logger(__name__)
|
||||
assert not utils.matches_filename('.', ('.*',), '', logger)
|
||||
assert not utils.matches_filename('..', ('.*',), '', logger)
|
||||
assert not utils.matches_filename(".", (".*",), "", logger)
|
||||
assert not utils.matches_filename("..", (".*",), "", logger)
|
||||
|
||||
|
||||
def test_stdin_get_value_crlf():
|
||||
"""Ensure that stdin is normalized from crlf to lf."""
|
||||
stdin = io.TextIOWrapper(io.BytesIO(b'1\r\n2\r\n'), 'UTF-8')
|
||||
with mock.patch.object(sys, 'stdin', stdin):
|
||||
assert utils.stdin_get_value.__wrapped__() == '1\n2\n'
|
||||
stdin = io.TextIOWrapper(io.BytesIO(b"1\r\n2\r\n"), "UTF-8")
|
||||
with mock.patch.object(sys, "stdin", stdin):
|
||||
assert utils.stdin_get_value.__wrapped__() == "1\n2\n"
|
||||
|
||||
|
||||
def test_stdin_unknown_coding_token():
|
||||
"""Ensure we produce source even for unknown encodings."""
|
||||
stdin = io.TextIOWrapper(io.BytesIO(b'# coding: unknown\n'), 'UTF-8')
|
||||
with mock.patch.object(sys, 'stdin', stdin):
|
||||
assert utils.stdin_get_value.__wrapped__() == '# coding: unknown\n'
|
||||
stdin = io.TextIOWrapper(io.BytesIO(b"# coding: unknown\n"), "UTF-8")
|
||||
with mock.patch.object(sys, "stdin", stdin):
|
||||
assert utils.stdin_get_value.__wrapped__() == "# coding: unknown\n"
|
||||
|
|
|
|||
|
|
@ -6,61 +6,74 @@ import pytest
|
|||
from flake8 import style_guide
|
||||
|
||||
|
||||
@pytest.mark.parametrize('error_code,physical_line,expected_result', [
|
||||
('E111', 'a = 1', False),
|
||||
('E121', 'a = 1 # noqa: E111', False),
|
||||
('E121', 'a = 1 # noqa: E111,W123,F821', False),
|
||||
('E111', 'a = 1 # noqa: E111,W123,F821', True),
|
||||
('W123', 'a = 1 # noqa: E111,W123,F821', True),
|
||||
('W123', 'a = 1 # noqa: E111, W123,F821', True),
|
||||
('E111', 'a = 1 # noqa: E11,W123,F821', True),
|
||||
('E121', 'a = 1 # noqa:E111,W123,F821', False),
|
||||
('E111', 'a = 1 # noqa:E111,W123,F821', True),
|
||||
('W123', 'a = 1 # noqa:E111,W123,F821', True),
|
||||
('W123', 'a = 1 # noqa:E111, W123,F821', True),
|
||||
('E111', 'a = 1 # noqa:E11,W123,F821', True),
|
||||
('E111', 'a = 1 # noqa, analysis:ignore', True),
|
||||
('E111', 'a = 1 # noqa analysis:ignore', True),
|
||||
('E111', 'a = 1 # noqa - We do not care', True),
|
||||
('E111', 'a = 1 # noqa: We do not care', True),
|
||||
('E111', 'a = 1 # noqa:We do not care', True),
|
||||
('ABC123', 'a = 1 # noqa: ABC123', True),
|
||||
('E111', 'a = 1 # noqa: ABC123', False),
|
||||
('ABC123', 'a = 1 # noqa: ABC124', False),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"error_code,physical_line,expected_result",
|
||||
[
|
||||
("E111", "a = 1", False),
|
||||
("E121", "a = 1 # noqa: E111", False),
|
||||
("E121", "a = 1 # noqa: E111,W123,F821", False),
|
||||
("E111", "a = 1 # noqa: E111,W123,F821", True),
|
||||
("W123", "a = 1 # noqa: E111,W123,F821", True),
|
||||
("W123", "a = 1 # noqa: E111, W123,F821", True),
|
||||
("E111", "a = 1 # noqa: E11,W123,F821", True),
|
||||
("E121", "a = 1 # noqa:E111,W123,F821", False),
|
||||
("E111", "a = 1 # noqa:E111,W123,F821", True),
|
||||
("W123", "a = 1 # noqa:E111,W123,F821", True),
|
||||
("W123", "a = 1 # noqa:E111, W123,F821", True),
|
||||
("E111", "a = 1 # noqa:E11,W123,F821", True),
|
||||
("E111", "a = 1 # noqa, analysis:ignore", True),
|
||||
("E111", "a = 1 # noqa analysis:ignore", True),
|
||||
("E111", "a = 1 # noqa - We do not care", True),
|
||||
("E111", "a = 1 # noqa: We do not care", True),
|
||||
("E111", "a = 1 # noqa:We do not care", True),
|
||||
("ABC123", "a = 1 # noqa: ABC123", True),
|
||||
("E111", "a = 1 # noqa: ABC123", False),
|
||||
("ABC123", "a = 1 # noqa: ABC124", False),
|
||||
],
|
||||
)
|
||||
def test_is_inline_ignored(error_code, physical_line, expected_result):
|
||||
"""Verify that we detect inline usage of ``# noqa``."""
|
||||
error = style_guide.Violation(
|
||||
error_code, 'filename.py', 1, 1, 'error text', None)
|
||||
error_code, "filename.py", 1, 1, "error text", None
|
||||
)
|
||||
# We want `None` to be passed as the physical line so we actually use our
|
||||
# monkey-patched linecache.getline value.
|
||||
|
||||
with mock.patch('linecache.getline', return_value=physical_line):
|
||||
with mock.patch("linecache.getline", return_value=physical_line):
|
||||
assert error.is_inline_ignored(False) is expected_result
|
||||
|
||||
|
||||
def test_disable_is_inline_ignored():
|
||||
"""Verify that is_inline_ignored exits immediately if disabling NoQA."""
|
||||
error = style_guide.Violation(
|
||||
'E121', 'filename.py', 1, 1, 'error text', 'line')
|
||||
"E121", "filename.py", 1, 1, "error text", "line"
|
||||
)
|
||||
|
||||
with mock.patch('linecache.getline') as getline:
|
||||
with mock.patch("linecache.getline") as getline:
|
||||
assert error.is_inline_ignored(True) is False
|
||||
|
||||
assert getline.called is False
|
||||
|
||||
|
||||
@pytest.mark.parametrize('violation_file,violation_line,diff,expected', [
|
||||
('file.py', 10, {}, True),
|
||||
('file.py', 1, {'file.py': range(1, 2)}, True),
|
||||
('file.py', 10, {'file.py': range(1, 2)}, False),
|
||||
('file.py', 1, {'other.py': range(1, 2)}, False),
|
||||
('file.py', 10, {'other.py': range(1, 2)}, False),
|
||||
])
|
||||
@pytest.mark.parametrize(
|
||||
"violation_file,violation_line,diff,expected",
|
||||
[
|
||||
("file.py", 10, {}, True),
|
||||
("file.py", 1, {"file.py": range(1, 2)}, True),
|
||||
("file.py", 10, {"file.py": range(1, 2)}, False),
|
||||
("file.py", 1, {"other.py": range(1, 2)}, False),
|
||||
("file.py", 10, {"other.py": range(1, 2)}, False),
|
||||
],
|
||||
)
|
||||
def test_violation_is_in_diff(violation_file, violation_line, diff, expected):
|
||||
"""Verify that we find violations within a diff."""
|
||||
violation = style_guide.Violation(
|
||||
'E001', violation_file, violation_line, 1, 'warning', 'line',
|
||||
"E001",
|
||||
violation_file,
|
||||
violation_line,
|
||||
1,
|
||||
"warning",
|
||||
"line",
|
||||
)
|
||||
|
||||
assert violation.is_in(diff) is expected
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue