python: Normalize quotes with Black.

Signed-off-by: Anders Kaseorg <anders@zulip.com>
This commit is contained in:
Anders Kaseorg
2021-02-11 23:20:45 -08:00
committed by Tim Abbott
parent 11741543da
commit 6e4c3e41dc
989 changed files with 32792 additions and 32792 deletions

View File

@@ -40,7 +40,7 @@ def tokenize(text: str) -> List[Token]:
def advance(n: int) -> None:
for _ in range(n):
state.i += 1
if state.i >= 0 and text[state.i - 1] == '\n':
if state.i >= 0 and text[state.i - 1] == "\n":
state.line += 1
state.col = 1
else:
@@ -94,22 +94,22 @@ def tokenize(text: str) -> List[Token]:
if looking_at_htmlcomment():
s = get_html_comment(text, state.i)
tag = s[4:-3]
kind = 'html_comment'
kind = "html_comment"
elif looking_at_handlebarcomment():
s = get_handlebar_comment(text, state.i)
tag = s[3:-2]
kind = 'handlebar_comment'
kind = "handlebar_comment"
elif looking_at_djangocomment():
s = get_django_comment(text, state.i)
tag = s[2:-2]
kind = 'django_comment'
kind = "django_comment"
elif looking_at_handlebarpartial():
s = get_handlebar_partial(text, state.i)
tag = s[9:-2]
kind = 'handlebars_singleton'
kind = "handlebars_singleton"
elif looking_at_html_start():
s = get_html_tag(text, state.i)
if s.endswith('/>'):
if s.endswith("/>"):
end_offset = -2
else:
end_offset = -1
@@ -121,42 +121,42 @@ def tokenize(text: str) -> List[Token]:
tag = tag_parts[0]
if is_special_html_tag(s, tag):
kind = 'html_special'
kind = "html_special"
elif is_self_closing_html_tag(s, tag):
kind = 'html_singleton'
kind = "html_singleton"
else:
kind = 'html_start'
kind = "html_start"
elif looking_at_html_end():
s = get_html_tag(text, state.i)
tag = s[2:-1]
kind = 'html_end'
kind = "html_end"
elif looking_at_handlebars_start():
s = get_handlebars_tag(text, state.i)
tag = s[3:-2].split()[0]
kind = 'handlebars_start'
kind = "handlebars_start"
elif looking_at_handlebars_end():
s = get_handlebars_tag(text, state.i)
tag = s[3:-2]
kind = 'handlebars_end'
kind = "handlebars_end"
elif looking_at_django_start():
s = get_django_tag(text, state.i)
tag = s[3:-2].split()[0]
kind = 'django_start'
kind = "django_start"
if s[-3] == '-':
kind = 'jinja2_whitespace_stripped_start'
if s[-3] == "-":
kind = "jinja2_whitespace_stripped_start"
elif looking_at_django_end():
s = get_django_tag(text, state.i)
tag = s[6:-3]
kind = 'django_end'
kind = "django_end"
elif looking_at_jinja2_end_whitespace_stripped():
s = get_django_tag(text, state.i)
tag = s[7:-3]
kind = 'jinja2_whitespace_stripped_end'
kind = "jinja2_whitespace_stripped_end"
elif looking_at_jinja2_start_whitespace_stripped_type2():
s = get_django_tag(text, state.i, stripped=True)
tag = s[3:-3].split()[0]
kind = 'jinja2_whitespace_stripped_type2_start'
kind = "jinja2_whitespace_stripped_type2_start"
else:
advance(1)
continue
@@ -165,7 +165,7 @@ def tokenize(text: str) -> List[Token]:
f'''{e.message} at Line {state.line} Col {state.col}:"{e.line_content}"''',
)
line_span = len(s.split('\n'))
line_span = len(s.split("\n"))
token = Token(
kind=kind,
s=s,
@@ -180,7 +180,7 @@ def tokenize(text: str) -> List[Token]:
def add_pseudo_end_token(kind: str) -> None:
token = Token(
kind=kind,
s='</' + tag + '>',
s="</" + tag + ">",
tag=tag,
line=state.line,
col=state.col,
@@ -188,15 +188,15 @@ def tokenize(text: str) -> List[Token]:
)
tokens.append(token)
if kind == 'html_singleton':
if kind == "html_singleton":
# Here we insert a Pseudo html_singleton_end tag so as to have
# ease of detection of end of singleton html tags which might be
# needed in some cases as with our html pretty printer.
add_pseudo_end_token('html_singleton_end')
if kind == 'handlebars_singleton':
add_pseudo_end_token("html_singleton_end")
if kind == "handlebars_singleton":
# We insert a pseudo handlbar end tag for singleton cases of
# handlebars like the partials. This helps in indenting multi line partials.
add_pseudo_end_token('handlebars_singleton_end')
add_pseudo_end_token("handlebars_singleton_end")
return tokens
@@ -207,7 +207,7 @@ def validate(
assert fn or text
if fn is None:
fn = '<in memory file>'
fn = "<in memory file>"
if text is None:
with open(fn) as f:
@@ -217,9 +217,9 @@ def validate(
tokens = tokenize(text)
except FormattedException as e:
raise TemplateParserException(
f'''
f"""
fn: {fn}
{e}'''
{e}"""
)
class State:
@@ -229,20 +229,20 @@ def validate(
def no_start_tag(token: Token) -> None:
raise TemplateParserException(
f'''
f"""
No start tag
fn: {fn}
end tag:
{token.tag}
line {token.line}, col {token.col}
'''
"""
)
state = State(no_start_tag)
def start_tag_matcher(start_token: Token) -> None:
state.depth += 1
start_tag = start_token.tag.strip('~')
start_tag = start_token.tag.strip("~")
start_line = start_token.line
start_col = start_token.col
@@ -250,26 +250,26 @@ def validate(
def f(end_token: Token) -> None:
end_tag = end_token.tag.strip('~')
end_tag = end_token.tag.strip("~")
end_line = end_token.line
end_col = end_token.col
if start_tag == 'a':
if start_tag == "a":
max_lines = 3
else:
max_lines = 1
problem = None
if (start_tag == 'code') and (end_line == start_line + 1):
problem = 'Code tag is split across two lines.'
if (start_tag == "code") and (end_line == start_line + 1):
problem = "Code tag is split across two lines."
if start_tag != end_tag:
problem = 'Mismatched tag.'
problem = "Mismatched tag."
elif check_indent and (end_line > start_line + max_lines):
if end_col != start_col:
problem = 'Bad indentation.'
problem = "Bad indentation."
if problem:
raise TemplateParserException(
f'''
f"""
fn: {fn}
{problem}
start:
@@ -278,7 +278,7 @@ def validate(
end tag:
{end_tag}
line {end_line}, col {end_col}
'''
"""
)
state.matcher = old_matcher
state.depth -= 1
@@ -289,62 +289,62 @@ def validate(
kind = token.kind
tag = token.tag
if kind == 'html_start':
if kind == "html_start":
start_tag_matcher(token)
elif kind == 'html_end':
elif kind == "html_end":
state.matcher(token)
elif kind == 'handlebars_start':
elif kind == "handlebars_start":
start_tag_matcher(token)
elif kind == 'handlebars_end':
elif kind == "handlebars_end":
state.matcher(token)
elif kind in {
'django_start',
'jinja2_whitespace_stripped_start',
'jinja2_whitespace_stripped_type2_start',
"django_start",
"jinja2_whitespace_stripped_start",
"jinja2_whitespace_stripped_type2_start",
}:
if is_django_block_tag(tag):
start_tag_matcher(token)
elif kind in {'django_end', 'jinja2_whitespace_stripped_end'}:
elif kind in {"django_end", "jinja2_whitespace_stripped_end"}:
state.matcher(token)
if state.depth != 0:
raise TemplateParserException('Missing end tag')
raise TemplateParserException("Missing end tag")
def is_special_html_tag(s: str, tag: str) -> bool:
return tag in ['link', 'meta', '!DOCTYPE']
return tag in ["link", "meta", "!DOCTYPE"]
OPTIONAL_CLOSING_TAGS = [
'circle',
'img',
'input',
'path',
'polygon',
'stop',
"circle",
"img",
"input",
"path",
"polygon",
"stop",
]
def is_self_closing_html_tag(s: str, tag: str) -> bool:
if s.endswith('/>'):
if s.endswith("/>"):
if tag in OPTIONAL_CLOSING_TAGS:
return True
raise TokenizationException('Singleton tag not allowed', tag)
raise TokenizationException("Singleton tag not allowed", tag)
self_closing_tag = tag in [
'area',
'base',
'br',
'col',
'embed',
'hr',
'img',
'input',
'param',
'source',
'track',
'wbr',
"area",
"base",
"br",
"col",
"embed",
"hr",
"img",
"input",
"param",
"source",
"track",
"wbr",
]
if self_closing_tag:
return True
@@ -353,26 +353,26 @@ def is_self_closing_html_tag(s: str, tag: str) -> bool:
def is_django_block_tag(tag: str) -> bool:
return tag in [
'autoescape',
'block',
'comment',
'for',
'if',
'ifequal',
'macro',
'verbatim',
'blocktrans',
'trans',
'raw',
'with',
"autoescape",
"block",
"comment",
"for",
"if",
"ifequal",
"macro",
"verbatim",
"blocktrans",
"trans",
"raw",
"with",
]
def get_handlebars_tag(text: str, i: int) -> str:
end = i + 2
while end < len(text) - 1 and text[end] != '}':
while end < len(text) - 1 and text[end] != "}":
end += 1
if text[end] != '}' or text[end + 1] != '}':
if text[end] != "}" or text[end + 1] != "}":
raise TokenizationException('Tag missing "}}"', text[i : end + 2])
s = text[i : end + 2]
return s
@@ -382,9 +382,9 @@ def get_django_tag(text: str, i: int, stripped: bool = False) -> str:
end = i + 2
if stripped:
end += 1
while end < len(text) - 1 and text[end] != '%':
while end < len(text) - 1 and text[end] != "%":
end += 1
if text[end] != '%' or text[end + 1] != '}':
if text[end] != "%" or text[end + 1] != "}":
raise TokenizationException('Tag missing "%}"', text[i : end + 2])
s = text[i : end + 2]
return s
@@ -394,18 +394,18 @@ def get_html_tag(text: str, i: int) -> str:
quote_count = 0
end = i + 1
unclosed_end = 0
while end < len(text) and (text[end] != '>' or quote_count % 2 != 0 and text[end] != '<'):
while end < len(text) and (text[end] != ">" or quote_count % 2 != 0 and text[end] != "<"):
if text[end] == '"':
quote_count += 1
if not unclosed_end and text[end] == '<':
if not unclosed_end and text[end] == "<":
unclosed_end = end
end += 1
if quote_count % 2 != 0:
if unclosed_end:
raise TokenizationException('Unbalanced Quotes', text[i:unclosed_end])
raise TokenizationException("Unbalanced Quotes", text[i:unclosed_end])
else:
raise TokenizationException('Unbalanced Quotes', text[i : end + 1])
if end == len(text) or text[end] != '>':
raise TokenizationException("Unbalanced Quotes", text[i : end + 1])
if end == len(text) or text[end] != ">":
raise TokenizationException('Tag missing ">"', text[i : end + 1])
s = text[i : end + 1]
return s
@@ -415,45 +415,45 @@ def get_html_comment(text: str, i: int) -> str:
end = i + 7
unclosed_end = 0
while end <= len(text):
if text[end - 3 : end] == '-->':
if text[end - 3 : end] == "-->":
return text[i:end]
if not unclosed_end and text[end] == '<':
if not unclosed_end and text[end] == "<":
unclosed_end = end
end += 1
raise TokenizationException('Unclosed comment', text[i:unclosed_end])
raise TokenizationException("Unclosed comment", text[i:unclosed_end])
def get_handlebar_comment(text: str, i: int) -> str:
end = i + 5
unclosed_end = 0
while end <= len(text):
if text[end - 2 : end] == '}}':
if text[end - 2 : end] == "}}":
return text[i:end]
if not unclosed_end and text[end] == '<':
if not unclosed_end and text[end] == "<":
unclosed_end = end
end += 1
raise TokenizationException('Unclosed comment', text[i:unclosed_end])
raise TokenizationException("Unclosed comment", text[i:unclosed_end])
def get_django_comment(text: str, i: int) -> str:
end = i + 4
unclosed_end = 0
while end <= len(text):
if text[end - 2 : end] == '#}':
if text[end - 2 : end] == "#}":
return text[i:end]
if not unclosed_end and text[end] == '<':
if not unclosed_end and text[end] == "<":
unclosed_end = end
end += 1
raise TokenizationException('Unclosed comment', text[i:unclosed_end])
raise TokenizationException("Unclosed comment", text[i:unclosed_end])
def get_handlebar_partial(text: str, i: int) -> str:
end = i + 10
unclosed_end = 0
while end <= len(text):
if text[end - 2 : end] == '}}':
if text[end - 2 : end] == "}}":
return text[i:end]
if not unclosed_end and text[end] == '<':
if not unclosed_end and text[end] == "<":
unclosed_end = end
end += 1
raise TokenizationException('Unclosed partial', text[i:unclosed_end])
raise TokenizationException("Unclosed partial", text[i:unclosed_end])