mirror of
				https://github.com/9001/copyparty.git
				synced 2025-11-03 21:43:12 +00:00 
			
		
		
		
	`tokenize.FSTRING_MIDDLE` was introduced, changing the
representation of `f"x{{y"` from `STRING(f"x{{y")` to:
* `FSTRING_START('f"')`
* `FSTRING_MIDDLE('x{')`
* `FSTRING_MIDDLE('y')`
* `FSTRING_END('"')`
each literal `{` (encoded as `{{` in the input) now appears as a
single `{` as the final character of its `FSTRING_MIDDLE`, with
additional consecutive `FSTRING_MIDDLE` tokens if necessary
regular interpolating `{` are encoded as separate `OP` tokens
the fact that the literal `{` is encoded as a single `{` instead
of `{{` breaks the assumption that the string-value of each token
maps directly to the original code
fix this by replacing `{` with `{{` and `}` with `}}` in
`FSTRING_MIDDLE` tokens, and not adding whitespace after
`FSTRING_MIDDLE` tokens
		
	
		
			
				
	
	
		
			108 lines
		
	
	
		
			2.8 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			108 lines
		
	
	
		
			2.8 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
#!/usr/bin/env python3
 | 
						|
# coding: utf-8
 | 
						|
from __future__ import print_function, unicode_literals
 | 
						|
 | 
						|
import io
 | 
						|
import os
 | 
						|
import sys
 | 
						|
import tokenize
 | 
						|
 | 
						|
 | 
						|
try:
 | 
						|
    FSTRING_MIDDLE = tokenize.FSTRING_MIDDLE
 | 
						|
except:
 | 
						|
    FSTRING_MIDDLE = -9001
 | 
						|
 | 
						|
 | 
						|
def uncomment(fpath):
 | 
						|
    """modified https://stackoverflow.com/a/62074206"""
 | 
						|
 | 
						|
    print(".", end="", flush=True)
 | 
						|
    with open(fpath, "rb") as f:
 | 
						|
        orig = f.read().decode("utf-8")
 | 
						|
 | 
						|
    out = ""
 | 
						|
    io_obj = io.StringIO(orig)
 | 
						|
    prev_toktype = tokenize.INDENT
 | 
						|
    last_lineno = -1
 | 
						|
    last_col = 0
 | 
						|
    code = False
 | 
						|
    for tok in tokenize.generate_tokens(io_obj.readline):
 | 
						|
        # print(repr(tok))
 | 
						|
        token_type = tok[0]
 | 
						|
        token_string = tok[1]
 | 
						|
        start_line, start_col = tok[2]
 | 
						|
        end_line, end_col = tok[3]
 | 
						|
 | 
						|
        if start_line > last_lineno:
 | 
						|
            last_col = 0
 | 
						|
 | 
						|
        if start_col > last_col and prev_toktype != FSTRING_MIDDLE:
 | 
						|
            out += " " * (start_col - last_col)
 | 
						|
 | 
						|
        is_legalese = (
 | 
						|
            "copyright" in token_string.lower() or "license" in token_string.lower()
 | 
						|
        )
 | 
						|
 | 
						|
        if token_type == tokenize.STRING:
 | 
						|
            if (
 | 
						|
                prev_toktype != tokenize.INDENT
 | 
						|
                and prev_toktype != tokenize.NEWLINE
 | 
						|
                and start_col > 0
 | 
						|
                or is_legalese
 | 
						|
            ):
 | 
						|
                out += token_string
 | 
						|
            else:
 | 
						|
                out += '"a"'
 | 
						|
        elif token_type == FSTRING_MIDDLE:
 | 
						|
            out += token_string.replace(r"{", r"{{").replace(r"}", r"}}")
 | 
						|
            if not code and token_string.strip():
 | 
						|
                code = True
 | 
						|
        elif token_type != tokenize.COMMENT:
 | 
						|
            out += token_string
 | 
						|
            if not code and token_string.strip():
 | 
						|
                code = True
 | 
						|
        elif is_legalese or (not start_col and not code):
 | 
						|
            out += token_string
 | 
						|
        else:
 | 
						|
            if out.rstrip(" ").endswith("\n"):
 | 
						|
                out = out.rstrip() + "\n"
 | 
						|
            else:
 | 
						|
                out = out.rstrip()
 | 
						|
 | 
						|
        prev_toktype = token_type
 | 
						|
        last_lineno = end_line
 | 
						|
        last_col = end_col
 | 
						|
 | 
						|
    # out = "\n".join(x for x in out.splitlines() if x.strip())
 | 
						|
 | 
						|
    with open(fpath, "wb") as f:
 | 
						|
        f.write(out.encode("utf-8"))
 | 
						|
 | 
						|
 | 
						|
def main():
 | 
						|
    print("uncommenting", end="", flush=True)
 | 
						|
    try:
 | 
						|
        if sys.argv[1] == "1":
 | 
						|
            sys.argv.remove("1")
 | 
						|
            raise Exception("disabled")
 | 
						|
 | 
						|
        import multiprocessing as mp
 | 
						|
 | 
						|
        mp.set_start_method("spawn", True)
 | 
						|
        with mp.Pool(os.cpu_count()) as pool:
 | 
						|
            pool.map(uncomment, sys.argv[1:])
 | 
						|
    except Exception as ex:
 | 
						|
        print("\nnon-mp fallback due to {}\n".format(ex))
 | 
						|
        for f in sys.argv[1:]:
 | 
						|
            try:
 | 
						|
                uncomment(f)
 | 
						|
            except Exception as ex:
 | 
						|
                print("uncomment failed: [%s] %s" % (f, repr(ex)))
 | 
						|
 | 
						|
    print("k")
 | 
						|
 | 
						|
 | 
						|
if __name__ == "__main__":
 | 
						|
    main()
 |