mirror of
				https://github.com/zulip/zulip.git
				synced 2025-11-04 05:53:43 +00:00 
			
		
		
		
	python: Normalize quotes with Black.
Signed-off-by: Anders Kaseorg <anders@zulip.com>
This commit is contained in:
		
				
					committed by
					
						
						Tim Abbott
					
				
			
			
				
	
			
			
			
						parent
						
							11741543da
						
					
				
				
					commit
					6e4c3e41dc
				
			@@ -40,7 +40,7 @@ def tokenize(text: str) -> List[Token]:
 | 
			
		||||
    def advance(n: int) -> None:
 | 
			
		||||
        for _ in range(n):
 | 
			
		||||
            state.i += 1
 | 
			
		||||
            if state.i >= 0 and text[state.i - 1] == '\n':
 | 
			
		||||
            if state.i >= 0 and text[state.i - 1] == "\n":
 | 
			
		||||
                state.line += 1
 | 
			
		||||
                state.col = 1
 | 
			
		||||
            else:
 | 
			
		||||
@@ -94,22 +94,22 @@ def tokenize(text: str) -> List[Token]:
 | 
			
		||||
            if looking_at_htmlcomment():
 | 
			
		||||
                s = get_html_comment(text, state.i)
 | 
			
		||||
                tag = s[4:-3]
 | 
			
		||||
                kind = 'html_comment'
 | 
			
		||||
                kind = "html_comment"
 | 
			
		||||
            elif looking_at_handlebarcomment():
 | 
			
		||||
                s = get_handlebar_comment(text, state.i)
 | 
			
		||||
                tag = s[3:-2]
 | 
			
		||||
                kind = 'handlebar_comment'
 | 
			
		||||
                kind = "handlebar_comment"
 | 
			
		||||
            elif looking_at_djangocomment():
 | 
			
		||||
                s = get_django_comment(text, state.i)
 | 
			
		||||
                tag = s[2:-2]
 | 
			
		||||
                kind = 'django_comment'
 | 
			
		||||
                kind = "django_comment"
 | 
			
		||||
            elif looking_at_handlebarpartial():
 | 
			
		||||
                s = get_handlebar_partial(text, state.i)
 | 
			
		||||
                tag = s[9:-2]
 | 
			
		||||
                kind = 'handlebars_singleton'
 | 
			
		||||
                kind = "handlebars_singleton"
 | 
			
		||||
            elif looking_at_html_start():
 | 
			
		||||
                s = get_html_tag(text, state.i)
 | 
			
		||||
                if s.endswith('/>'):
 | 
			
		||||
                if s.endswith("/>"):
 | 
			
		||||
                    end_offset = -2
 | 
			
		||||
                else:
 | 
			
		||||
                    end_offset = -1
 | 
			
		||||
@@ -121,42 +121,42 @@ def tokenize(text: str) -> List[Token]:
 | 
			
		||||
                tag = tag_parts[0]
 | 
			
		||||
 | 
			
		||||
                if is_special_html_tag(s, tag):
 | 
			
		||||
                    kind = 'html_special'
 | 
			
		||||
                    kind = "html_special"
 | 
			
		||||
                elif is_self_closing_html_tag(s, tag):
 | 
			
		||||
                    kind = 'html_singleton'
 | 
			
		||||
                    kind = "html_singleton"
 | 
			
		||||
                else:
 | 
			
		||||
                    kind = 'html_start'
 | 
			
		||||
                    kind = "html_start"
 | 
			
		||||
            elif looking_at_html_end():
 | 
			
		||||
                s = get_html_tag(text, state.i)
 | 
			
		||||
                tag = s[2:-1]
 | 
			
		||||
                kind = 'html_end'
 | 
			
		||||
                kind = "html_end"
 | 
			
		||||
            elif looking_at_handlebars_start():
 | 
			
		||||
                s = get_handlebars_tag(text, state.i)
 | 
			
		||||
                tag = s[3:-2].split()[0]
 | 
			
		||||
                kind = 'handlebars_start'
 | 
			
		||||
                kind = "handlebars_start"
 | 
			
		||||
            elif looking_at_handlebars_end():
 | 
			
		||||
                s = get_handlebars_tag(text, state.i)
 | 
			
		||||
                tag = s[3:-2]
 | 
			
		||||
                kind = 'handlebars_end'
 | 
			
		||||
                kind = "handlebars_end"
 | 
			
		||||
            elif looking_at_django_start():
 | 
			
		||||
                s = get_django_tag(text, state.i)
 | 
			
		||||
                tag = s[3:-2].split()[0]
 | 
			
		||||
                kind = 'django_start'
 | 
			
		||||
                kind = "django_start"
 | 
			
		||||
 | 
			
		||||
                if s[-3] == '-':
 | 
			
		||||
                    kind = 'jinja2_whitespace_stripped_start'
 | 
			
		||||
                if s[-3] == "-":
 | 
			
		||||
                    kind = "jinja2_whitespace_stripped_start"
 | 
			
		||||
            elif looking_at_django_end():
 | 
			
		||||
                s = get_django_tag(text, state.i)
 | 
			
		||||
                tag = s[6:-3]
 | 
			
		||||
                kind = 'django_end'
 | 
			
		||||
                kind = "django_end"
 | 
			
		||||
            elif looking_at_jinja2_end_whitespace_stripped():
 | 
			
		||||
                s = get_django_tag(text, state.i)
 | 
			
		||||
                tag = s[7:-3]
 | 
			
		||||
                kind = 'jinja2_whitespace_stripped_end'
 | 
			
		||||
                kind = "jinja2_whitespace_stripped_end"
 | 
			
		||||
            elif looking_at_jinja2_start_whitespace_stripped_type2():
 | 
			
		||||
                s = get_django_tag(text, state.i, stripped=True)
 | 
			
		||||
                tag = s[3:-3].split()[0]
 | 
			
		||||
                kind = 'jinja2_whitespace_stripped_type2_start'
 | 
			
		||||
                kind = "jinja2_whitespace_stripped_type2_start"
 | 
			
		||||
            else:
 | 
			
		||||
                advance(1)
 | 
			
		||||
                continue
 | 
			
		||||
@@ -165,7 +165,7 @@ def tokenize(text: str) -> List[Token]:
 | 
			
		||||
                f'''{e.message} at Line {state.line} Col {state.col}:"{e.line_content}"''',
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        line_span = len(s.split('\n'))
 | 
			
		||||
        line_span = len(s.split("\n"))
 | 
			
		||||
        token = Token(
 | 
			
		||||
            kind=kind,
 | 
			
		||||
            s=s,
 | 
			
		||||
@@ -180,7 +180,7 @@ def tokenize(text: str) -> List[Token]:
 | 
			
		||||
        def add_pseudo_end_token(kind: str) -> None:
 | 
			
		||||
            token = Token(
 | 
			
		||||
                kind=kind,
 | 
			
		||||
                s='</' + tag + '>',
 | 
			
		||||
                s="</" + tag + ">",
 | 
			
		||||
                tag=tag,
 | 
			
		||||
                line=state.line,
 | 
			
		||||
                col=state.col,
 | 
			
		||||
@@ -188,15 +188,15 @@ def tokenize(text: str) -> List[Token]:
 | 
			
		||||
            )
 | 
			
		||||
            tokens.append(token)
 | 
			
		||||
 | 
			
		||||
        if kind == 'html_singleton':
 | 
			
		||||
        if kind == "html_singleton":
 | 
			
		||||
            # Here we insert a Pseudo html_singleton_end tag so as to have
 | 
			
		||||
            # ease of detection of end of singleton html tags which might be
 | 
			
		||||
            # needed in some cases as with our html pretty printer.
 | 
			
		||||
            add_pseudo_end_token('html_singleton_end')
 | 
			
		||||
        if kind == 'handlebars_singleton':
 | 
			
		||||
            add_pseudo_end_token("html_singleton_end")
 | 
			
		||||
        if kind == "handlebars_singleton":
 | 
			
		||||
            # We insert a pseudo handlbar end tag for singleton cases of
 | 
			
		||||
            # handlebars like the partials. This helps in indenting multi line partials.
 | 
			
		||||
            add_pseudo_end_token('handlebars_singleton_end')
 | 
			
		||||
            add_pseudo_end_token("handlebars_singleton_end")
 | 
			
		||||
 | 
			
		||||
    return tokens
 | 
			
		||||
 | 
			
		||||
@@ -207,7 +207,7 @@ def validate(
 | 
			
		||||
    assert fn or text
 | 
			
		||||
 | 
			
		||||
    if fn is None:
 | 
			
		||||
        fn = '<in memory file>'
 | 
			
		||||
        fn = "<in memory file>"
 | 
			
		||||
 | 
			
		||||
    if text is None:
 | 
			
		||||
        with open(fn) as f:
 | 
			
		||||
@@ -217,9 +217,9 @@ def validate(
 | 
			
		||||
        tokens = tokenize(text)
 | 
			
		||||
    except FormattedException as e:
 | 
			
		||||
        raise TemplateParserException(
 | 
			
		||||
            f'''
 | 
			
		||||
            f"""
 | 
			
		||||
            fn: {fn}
 | 
			
		||||
            {e}'''
 | 
			
		||||
            {e}"""
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    class State:
 | 
			
		||||
@@ -229,20 +229,20 @@ def validate(
 | 
			
		||||
 | 
			
		||||
    def no_start_tag(token: Token) -> None:
 | 
			
		||||
        raise TemplateParserException(
 | 
			
		||||
            f'''
 | 
			
		||||
            f"""
 | 
			
		||||
            No start tag
 | 
			
		||||
            fn: {fn}
 | 
			
		||||
            end tag:
 | 
			
		||||
                {token.tag}
 | 
			
		||||
                line {token.line}, col {token.col}
 | 
			
		||||
            '''
 | 
			
		||||
            """
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    state = State(no_start_tag)
 | 
			
		||||
 | 
			
		||||
    def start_tag_matcher(start_token: Token) -> None:
 | 
			
		||||
        state.depth += 1
 | 
			
		||||
        start_tag = start_token.tag.strip('~')
 | 
			
		||||
        start_tag = start_token.tag.strip("~")
 | 
			
		||||
        start_line = start_token.line
 | 
			
		||||
        start_col = start_token.col
 | 
			
		||||
 | 
			
		||||
@@ -250,26 +250,26 @@ def validate(
 | 
			
		||||
 | 
			
		||||
        def f(end_token: Token) -> None:
 | 
			
		||||
 | 
			
		||||
            end_tag = end_token.tag.strip('~')
 | 
			
		||||
            end_tag = end_token.tag.strip("~")
 | 
			
		||||
            end_line = end_token.line
 | 
			
		||||
            end_col = end_token.col
 | 
			
		||||
 | 
			
		||||
            if start_tag == 'a':
 | 
			
		||||
            if start_tag == "a":
 | 
			
		||||
                max_lines = 3
 | 
			
		||||
            else:
 | 
			
		||||
                max_lines = 1
 | 
			
		||||
 | 
			
		||||
            problem = None
 | 
			
		||||
            if (start_tag == 'code') and (end_line == start_line + 1):
 | 
			
		||||
                problem = 'Code tag is split across two lines.'
 | 
			
		||||
            if (start_tag == "code") and (end_line == start_line + 1):
 | 
			
		||||
                problem = "Code tag is split across two lines."
 | 
			
		||||
            if start_tag != end_tag:
 | 
			
		||||
                problem = 'Mismatched tag.'
 | 
			
		||||
                problem = "Mismatched tag."
 | 
			
		||||
            elif check_indent and (end_line > start_line + max_lines):
 | 
			
		||||
                if end_col != start_col:
 | 
			
		||||
                    problem = 'Bad indentation.'
 | 
			
		||||
                    problem = "Bad indentation."
 | 
			
		||||
            if problem:
 | 
			
		||||
                raise TemplateParserException(
 | 
			
		||||
                    f'''
 | 
			
		||||
                    f"""
 | 
			
		||||
                    fn: {fn}
 | 
			
		||||
                    {problem}
 | 
			
		||||
                    start:
 | 
			
		||||
@@ -278,7 +278,7 @@ def validate(
 | 
			
		||||
                    end tag:
 | 
			
		||||
                        {end_tag}
 | 
			
		||||
                        line {end_line}, col {end_col}
 | 
			
		||||
                    '''
 | 
			
		||||
                    """
 | 
			
		||||
                )
 | 
			
		||||
            state.matcher = old_matcher
 | 
			
		||||
            state.depth -= 1
 | 
			
		||||
@@ -289,62 +289,62 @@ def validate(
 | 
			
		||||
        kind = token.kind
 | 
			
		||||
        tag = token.tag
 | 
			
		||||
 | 
			
		||||
        if kind == 'html_start':
 | 
			
		||||
        if kind == "html_start":
 | 
			
		||||
            start_tag_matcher(token)
 | 
			
		||||
        elif kind == 'html_end':
 | 
			
		||||
        elif kind == "html_end":
 | 
			
		||||
            state.matcher(token)
 | 
			
		||||
 | 
			
		||||
        elif kind == 'handlebars_start':
 | 
			
		||||
        elif kind == "handlebars_start":
 | 
			
		||||
            start_tag_matcher(token)
 | 
			
		||||
        elif kind == 'handlebars_end':
 | 
			
		||||
        elif kind == "handlebars_end":
 | 
			
		||||
            state.matcher(token)
 | 
			
		||||
 | 
			
		||||
        elif kind in {
 | 
			
		||||
            'django_start',
 | 
			
		||||
            'jinja2_whitespace_stripped_start',
 | 
			
		||||
            'jinja2_whitespace_stripped_type2_start',
 | 
			
		||||
            "django_start",
 | 
			
		||||
            "jinja2_whitespace_stripped_start",
 | 
			
		||||
            "jinja2_whitespace_stripped_type2_start",
 | 
			
		||||
        }:
 | 
			
		||||
            if is_django_block_tag(tag):
 | 
			
		||||
                start_tag_matcher(token)
 | 
			
		||||
        elif kind in {'django_end', 'jinja2_whitespace_stripped_end'}:
 | 
			
		||||
        elif kind in {"django_end", "jinja2_whitespace_stripped_end"}:
 | 
			
		||||
            state.matcher(token)
 | 
			
		||||
 | 
			
		||||
    if state.depth != 0:
 | 
			
		||||
        raise TemplateParserException('Missing end tag')
 | 
			
		||||
        raise TemplateParserException("Missing end tag")
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def is_special_html_tag(s: str, tag: str) -> bool:
 | 
			
		||||
    return tag in ['link', 'meta', '!DOCTYPE']
 | 
			
		||||
    return tag in ["link", "meta", "!DOCTYPE"]
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
OPTIONAL_CLOSING_TAGS = [
 | 
			
		||||
    'circle',
 | 
			
		||||
    'img',
 | 
			
		||||
    'input',
 | 
			
		||||
    'path',
 | 
			
		||||
    'polygon',
 | 
			
		||||
    'stop',
 | 
			
		||||
    "circle",
 | 
			
		||||
    "img",
 | 
			
		||||
    "input",
 | 
			
		||||
    "path",
 | 
			
		||||
    "polygon",
 | 
			
		||||
    "stop",
 | 
			
		||||
]
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def is_self_closing_html_tag(s: str, tag: str) -> bool:
 | 
			
		||||
    if s.endswith('/>'):
 | 
			
		||||
    if s.endswith("/>"):
 | 
			
		||||
        if tag in OPTIONAL_CLOSING_TAGS:
 | 
			
		||||
            return True
 | 
			
		||||
        raise TokenizationException('Singleton tag not allowed', tag)
 | 
			
		||||
        raise TokenizationException("Singleton tag not allowed", tag)
 | 
			
		||||
    self_closing_tag = tag in [
 | 
			
		||||
        'area',
 | 
			
		||||
        'base',
 | 
			
		||||
        'br',
 | 
			
		||||
        'col',
 | 
			
		||||
        'embed',
 | 
			
		||||
        'hr',
 | 
			
		||||
        'img',
 | 
			
		||||
        'input',
 | 
			
		||||
        'param',
 | 
			
		||||
        'source',
 | 
			
		||||
        'track',
 | 
			
		||||
        'wbr',
 | 
			
		||||
        "area",
 | 
			
		||||
        "base",
 | 
			
		||||
        "br",
 | 
			
		||||
        "col",
 | 
			
		||||
        "embed",
 | 
			
		||||
        "hr",
 | 
			
		||||
        "img",
 | 
			
		||||
        "input",
 | 
			
		||||
        "param",
 | 
			
		||||
        "source",
 | 
			
		||||
        "track",
 | 
			
		||||
        "wbr",
 | 
			
		||||
    ]
 | 
			
		||||
    if self_closing_tag:
 | 
			
		||||
        return True
 | 
			
		||||
@@ -353,26 +353,26 @@ def is_self_closing_html_tag(s: str, tag: str) -> bool:
 | 
			
		||||
 | 
			
		||||
def is_django_block_tag(tag: str) -> bool:
 | 
			
		||||
    return tag in [
 | 
			
		||||
        'autoescape',
 | 
			
		||||
        'block',
 | 
			
		||||
        'comment',
 | 
			
		||||
        'for',
 | 
			
		||||
        'if',
 | 
			
		||||
        'ifequal',
 | 
			
		||||
        'macro',
 | 
			
		||||
        'verbatim',
 | 
			
		||||
        'blocktrans',
 | 
			
		||||
        'trans',
 | 
			
		||||
        'raw',
 | 
			
		||||
        'with',
 | 
			
		||||
        "autoescape",
 | 
			
		||||
        "block",
 | 
			
		||||
        "comment",
 | 
			
		||||
        "for",
 | 
			
		||||
        "if",
 | 
			
		||||
        "ifequal",
 | 
			
		||||
        "macro",
 | 
			
		||||
        "verbatim",
 | 
			
		||||
        "blocktrans",
 | 
			
		||||
        "trans",
 | 
			
		||||
        "raw",
 | 
			
		||||
        "with",
 | 
			
		||||
    ]
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def get_handlebars_tag(text: str, i: int) -> str:
 | 
			
		||||
    end = i + 2
 | 
			
		||||
    while end < len(text) - 1 and text[end] != '}':
 | 
			
		||||
    while end < len(text) - 1 and text[end] != "}":
 | 
			
		||||
        end += 1
 | 
			
		||||
    if text[end] != '}' or text[end + 1] != '}':
 | 
			
		||||
    if text[end] != "}" or text[end + 1] != "}":
 | 
			
		||||
        raise TokenizationException('Tag missing "}}"', text[i : end + 2])
 | 
			
		||||
    s = text[i : end + 2]
 | 
			
		||||
    return s
 | 
			
		||||
@@ -382,9 +382,9 @@ def get_django_tag(text: str, i: int, stripped: bool = False) -> str:
 | 
			
		||||
    end = i + 2
 | 
			
		||||
    if stripped:
 | 
			
		||||
        end += 1
 | 
			
		||||
    while end < len(text) - 1 and text[end] != '%':
 | 
			
		||||
    while end < len(text) - 1 and text[end] != "%":
 | 
			
		||||
        end += 1
 | 
			
		||||
    if text[end] != '%' or text[end + 1] != '}':
 | 
			
		||||
    if text[end] != "%" or text[end + 1] != "}":
 | 
			
		||||
        raise TokenizationException('Tag missing "%}"', text[i : end + 2])
 | 
			
		||||
    s = text[i : end + 2]
 | 
			
		||||
    return s
 | 
			
		||||
@@ -394,18 +394,18 @@ def get_html_tag(text: str, i: int) -> str:
 | 
			
		||||
    quote_count = 0
 | 
			
		||||
    end = i + 1
 | 
			
		||||
    unclosed_end = 0
 | 
			
		||||
    while end < len(text) and (text[end] != '>' or quote_count % 2 != 0 and text[end] != '<'):
 | 
			
		||||
    while end < len(text) and (text[end] != ">" or quote_count % 2 != 0 and text[end] != "<"):
 | 
			
		||||
        if text[end] == '"':
 | 
			
		||||
            quote_count += 1
 | 
			
		||||
        if not unclosed_end and text[end] == '<':
 | 
			
		||||
        if not unclosed_end and text[end] == "<":
 | 
			
		||||
            unclosed_end = end
 | 
			
		||||
        end += 1
 | 
			
		||||
    if quote_count % 2 != 0:
 | 
			
		||||
        if unclosed_end:
 | 
			
		||||
            raise TokenizationException('Unbalanced Quotes', text[i:unclosed_end])
 | 
			
		||||
            raise TokenizationException("Unbalanced Quotes", text[i:unclosed_end])
 | 
			
		||||
        else:
 | 
			
		||||
            raise TokenizationException('Unbalanced Quotes', text[i : end + 1])
 | 
			
		||||
    if end == len(text) or text[end] != '>':
 | 
			
		||||
            raise TokenizationException("Unbalanced Quotes", text[i : end + 1])
 | 
			
		||||
    if end == len(text) or text[end] != ">":
 | 
			
		||||
        raise TokenizationException('Tag missing ">"', text[i : end + 1])
 | 
			
		||||
    s = text[i : end + 1]
 | 
			
		||||
    return s
 | 
			
		||||
@@ -415,45 +415,45 @@ def get_html_comment(text: str, i: int) -> str:
 | 
			
		||||
    end = i + 7
 | 
			
		||||
    unclosed_end = 0
 | 
			
		||||
    while end <= len(text):
 | 
			
		||||
        if text[end - 3 : end] == '-->':
 | 
			
		||||
        if text[end - 3 : end] == "-->":
 | 
			
		||||
            return text[i:end]
 | 
			
		||||
        if not unclosed_end and text[end] == '<':
 | 
			
		||||
        if not unclosed_end and text[end] == "<":
 | 
			
		||||
            unclosed_end = end
 | 
			
		||||
        end += 1
 | 
			
		||||
    raise TokenizationException('Unclosed comment', text[i:unclosed_end])
 | 
			
		||||
    raise TokenizationException("Unclosed comment", text[i:unclosed_end])
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def get_handlebar_comment(text: str, i: int) -> str:
 | 
			
		||||
    end = i + 5
 | 
			
		||||
    unclosed_end = 0
 | 
			
		||||
    while end <= len(text):
 | 
			
		||||
        if text[end - 2 : end] == '}}':
 | 
			
		||||
        if text[end - 2 : end] == "}}":
 | 
			
		||||
            return text[i:end]
 | 
			
		||||
        if not unclosed_end and text[end] == '<':
 | 
			
		||||
        if not unclosed_end and text[end] == "<":
 | 
			
		||||
            unclosed_end = end
 | 
			
		||||
        end += 1
 | 
			
		||||
    raise TokenizationException('Unclosed comment', text[i:unclosed_end])
 | 
			
		||||
    raise TokenizationException("Unclosed comment", text[i:unclosed_end])
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def get_django_comment(text: str, i: int) -> str:
 | 
			
		||||
    end = i + 4
 | 
			
		||||
    unclosed_end = 0
 | 
			
		||||
    while end <= len(text):
 | 
			
		||||
        if text[end - 2 : end] == '#}':
 | 
			
		||||
        if text[end - 2 : end] == "#}":
 | 
			
		||||
            return text[i:end]
 | 
			
		||||
        if not unclosed_end and text[end] == '<':
 | 
			
		||||
        if not unclosed_end and text[end] == "<":
 | 
			
		||||
            unclosed_end = end
 | 
			
		||||
        end += 1
 | 
			
		||||
    raise TokenizationException('Unclosed comment', text[i:unclosed_end])
 | 
			
		||||
    raise TokenizationException("Unclosed comment", text[i:unclosed_end])
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def get_handlebar_partial(text: str, i: int) -> str:
 | 
			
		||||
    end = i + 10
 | 
			
		||||
    unclosed_end = 0
 | 
			
		||||
    while end <= len(text):
 | 
			
		||||
        if text[end - 2 : end] == '}}':
 | 
			
		||||
        if text[end - 2 : end] == "}}":
 | 
			
		||||
            return text[i:end]
 | 
			
		||||
        if not unclosed_end and text[end] == '<':
 | 
			
		||||
        if not unclosed_end and text[end] == "<":
 | 
			
		||||
            unclosed_end = end
 | 
			
		||||
        end += 1
 | 
			
		||||
    raise TokenizationException('Unclosed partial', text[i:unclosed_end])
 | 
			
		||||
    raise TokenizationException("Unclosed partial", text[i:unclosed_end])
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user