Skip to content

Commit

Permalink
reformat
Browse files Browse the repository at this point in the history
  • Loading branch information
JelleZijlstra committed May 3, 2023
1 parent 1c028c7 commit 3cbd3f0
Showing 1 changed file with 6 additions and 4 deletions.
10 changes: 6 additions & 4 deletions src/blib2to3/pgen2/tokenize.py
Original file line number Diff line number Diff line change
Expand Up @@ -191,7 +191,9 @@ class StopTokenizing(Exception):
Coord = Tuple[int, int]


def printtoken(type: int, token: Text, srow_col: Coord, erow_col: Coord, line: Text): # for testing
def printtoken(
type: int, token: Text, srow_col: Coord, erow_col: Coord, line: Text
): # for testing
(srow, scol) = srow_col
(erow, ecol) = erow_col
print(
Expand Down Expand Up @@ -232,7 +234,6 @@ def tokenize_loop(readline: Callable[[], Text], tokeneater: TokenEater) -> None:


class Untokenizer:

tokens: List[Text]
prev_row: int
prev_col: int
Expand Down Expand Up @@ -606,7 +607,9 @@ def generate_tokens(
or endprogs.get(token[1])
or endprogs.get(token[2])
)
assert maybe_endprog is not None, f"endprog not found for {token}"
assert (
maybe_endprog is not None
), f"endprog not found for {token}"
endprog = maybe_endprog
contstr, needcont = line[start:], 1
contline = line
Expand Down Expand Up @@ -635,7 +638,6 @@ def generate_tokens(

if token in ("def", "for"):
if stashed and stashed[0] == NAME and stashed[1] == "async":

if token == "def":
async_def = True
async_def_indent = indents[-1]
Expand Down

0 comments on commit 3cbd3f0

Please sign in to comment.