mirror of
https://github.com/PyCQA/flake8.git
synced 2026-03-29 10:36:53 +00:00
use tokenize.TokenInfo instead of _Token alias
This commit is contained in:
parent
01e8376094
commit
ec57d5e67c
4 changed files with 18 additions and 14 deletions
|
|
@ -275,7 +275,7 @@ def test_split_line(unsplit_line, expected_lines, default_options):
|
|||
],
|
||||
)
|
||||
|
||||
token = (1, unsplit_line, (0, 0), (0, 0), "")
|
||||
token = tokenize.TokenInfo(1, unsplit_line, (0, 0), (0, 0), "")
|
||||
actual_lines = list(file_processor.split_line(token))
|
||||
assert expected_lines == actual_lines
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue