mirror of
https://github.com/9001/copyparty.git
synced 2025-11-05 14:23:17 +00:00
Compare commits
115 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
44a78a7e21 | ||
|
|
6b75cbf747 | ||
|
|
e7b18ab9fe | ||
|
|
aa12830015 | ||
|
|
f156e00064 | ||
|
|
d53c212516 | ||
|
|
ca27f8587c | ||
|
|
88ce008e16 | ||
|
|
081d2cc5d7 | ||
|
|
60ac68d000 | ||
|
|
fbe656957d | ||
|
|
5534c78c17 | ||
|
|
a45a53fdce | ||
|
|
972a56e738 | ||
|
|
5e03b3ca38 | ||
|
|
1078d933b4 | ||
|
|
d6bf300d80 | ||
|
|
a359d64d44 | ||
|
|
22396e8c33 | ||
|
|
5ded5a4516 | ||
|
|
79c7639aaf | ||
|
|
5bbf875385 | ||
|
|
5e159432af | ||
|
|
1d6ae409f6 | ||
|
|
9d729d3d1a | ||
|
|
4dd5d4e1b7 | ||
|
|
acd8149479 | ||
|
|
b97a1088fa | ||
|
|
b77bed3324 | ||
|
|
a2b7c85a1f | ||
|
|
b28533f850 | ||
|
|
bd8c7e538a | ||
|
|
89e48cff24 | ||
|
|
ae90a7b7b6 | ||
|
|
6fc1be04da | ||
|
|
0061d29534 | ||
|
|
a891f34a93 | ||
|
|
d6a1e62a95 | ||
|
|
cda36ea8b4 | ||
|
|
909a76434a | ||
|
|
39348ef659 | ||
|
|
99d30edef3 | ||
|
|
b63ab15bf9 | ||
|
|
485cb4495c | ||
|
|
df018eb1f2 | ||
|
|
49aa47a9b8 | ||
|
|
7d20eb202a | ||
|
|
c533da9129 | ||
|
|
5cba31a814 | ||
|
|
1d824cb26c | ||
|
|
83b903d60e | ||
|
|
9c8ccabe8e | ||
|
|
b1f2c4e70d | ||
|
|
273ca0c8da | ||
|
|
d6f516b34f | ||
|
|
83127858ca | ||
|
|
d89329757e | ||
|
|
49ffec5320 | ||
|
|
2eaae2b66a | ||
|
|
ea4441e25c | ||
|
|
e5f34042f9 | ||
|
|
271096874a | ||
|
|
8efd780a72 | ||
|
|
41bcf7308d | ||
|
|
d102bb3199 | ||
|
|
d0bed95415 | ||
|
|
2528729971 | ||
|
|
292c18b3d0 | ||
|
|
0be7c5e2d8 | ||
|
|
eb5aaddba4 | ||
|
|
d8fd82bcb5 | ||
|
|
97be495861 | ||
|
|
8b53c159fc | ||
|
|
81e281f703 | ||
|
|
3948214050 | ||
|
|
c5e9a643e7 | ||
|
|
d25881d5c3 | ||
|
|
38d8d9733f | ||
|
|
118ebf668d | ||
|
|
a86f09fa46 | ||
|
|
dd4fb35c8f | ||
|
|
621eb4cf95 | ||
|
|
deea66ad0b | ||
|
|
bf99445377 | ||
|
|
7b54a63396 | ||
|
|
0fcb015f9a | ||
|
|
0a22b1ffb6 | ||
|
|
68cecc52ab | ||
|
|
53657ccfff | ||
|
|
96223fda01 | ||
|
|
374ff3433e | ||
|
|
5d63949e98 | ||
|
|
6b065d507d | ||
|
|
e79997498a | ||
|
|
f7ee02ec35 | ||
|
|
69dc433e1c | ||
|
|
c880cd848c | ||
|
|
5752b6db48 | ||
|
|
b36f905eab | ||
|
|
483dd527c6 | ||
|
|
e55678e28f | ||
|
|
3f4a8b9d6f | ||
|
|
02a856ecb4 | ||
|
|
4dff726310 | ||
|
|
cbc449036f | ||
|
|
8f53152220 | ||
|
|
bbb1e165d6 | ||
|
|
fed8d94885 | ||
|
|
58040cc0ed | ||
|
|
03d692db66 | ||
|
|
903f8e8453 | ||
|
|
405ae1308e | ||
|
|
8a0f583d71 | ||
|
|
b6d7017491 | ||
|
|
0f0217d203 |
17
.vscode/launch.json
vendored
17
.vscode/launch.json
vendored
@@ -16,12 +16,9 @@
|
|||||||
"-e2ts",
|
"-e2ts",
|
||||||
"-mtp",
|
"-mtp",
|
||||||
".bpm=f,bin/mtag/audio-bpm.py",
|
".bpm=f,bin/mtag/audio-bpm.py",
|
||||||
"-a",
|
"-aed:wark",
|
||||||
"ed:wark",
|
"-vsrv::r:aed:cnodupe",
|
||||||
"-v",
|
"-vdist:dist:r"
|
||||||
"srv::r:aed:cnodupe",
|
|
||||||
"-v",
|
|
||||||
"dist:dist:r"
|
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -43,5 +40,13 @@
|
|||||||
"${file}"
|
"${file}"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"name": "Python: Current File",
|
||||||
|
"type": "python",
|
||||||
|
"request": "launch",
|
||||||
|
"program": "${file}",
|
||||||
|
"console": "integratedTerminal",
|
||||||
|
"justMyCode": false
|
||||||
|
},
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
18
.vscode/launch.py
vendored
18
.vscode/launch.py
vendored
@@ -3,14 +3,16 @@
|
|||||||
# launches 10x faster than mspython debugpy
|
# launches 10x faster than mspython debugpy
|
||||||
# and is stoppable with ^C
|
# and is stoppable with ^C
|
||||||
|
|
||||||
|
import re
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
print(sys.executable)
|
||||||
|
|
||||||
import shlex
|
import shlex
|
||||||
|
|
||||||
sys.path.insert(0, os.getcwd())
|
|
||||||
|
|
||||||
import jstyleson
|
import jstyleson
|
||||||
from copyparty.__main__ import main as copyparty
|
import subprocess as sp
|
||||||
|
|
||||||
|
|
||||||
with open(".vscode/launch.json", "r", encoding="utf-8") as f:
|
with open(".vscode/launch.json", "r", encoding="utf-8") as f:
|
||||||
tj = f.read()
|
tj = f.read()
|
||||||
@@ -25,6 +27,14 @@ except:
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
argv = [os.path.expanduser(x) if x.startswith("~") else x for x in argv]
|
argv = [os.path.expanduser(x) if x.startswith("~") else x for x in argv]
|
||||||
|
|
||||||
|
if re.search(" -j ?[0-9]", " ".join(argv)):
|
||||||
|
argv = [sys.executable, "-m", "copyparty"] + argv
|
||||||
|
sp.check_call(argv)
|
||||||
|
else:
|
||||||
|
sys.path.insert(0, os.getcwd())
|
||||||
|
from copyparty.__main__ import main as copyparty
|
||||||
|
|
||||||
try:
|
try:
|
||||||
copyparty(["a"] + argv)
|
copyparty(["a"] + argv)
|
||||||
except SystemExit as ex:
|
except SystemExit as ex:
|
||||||
|
|||||||
160
README.md
160
README.md
@@ -10,10 +10,10 @@ turn your phone or raspi into a portable file server with resumable uploads/down
|
|||||||
|
|
||||||
* server runs on anything with `py2.7` or `py3.3+`
|
* server runs on anything with `py2.7` or `py3.3+`
|
||||||
* browse/upload with IE4 / netscape4.0 on win3.11 (heh)
|
* browse/upload with IE4 / netscape4.0 on win3.11 (heh)
|
||||||
* *resumable* uploads need `firefox 34+` / `chrome 37+` / `safari 7+`
|
* *resumable* uploads need `firefox 34+` / `chrome 41+` / `safari 7+` for full speed
|
||||||
* code standard: `black`
|
* code standard: `black`
|
||||||
|
|
||||||
📷 screenshots: [browser](#the-browser) // [upload](#uploading) // [md-viewer](#markdown-viewer) // [search](#searching) // [fsearch](#file-search) // [zip-DL](#zip-downloads) // [ie4](#browser-support)
|
📷 **screenshots:** [browser](#the-browser) // [upload](#uploading) // [thumbnails](#thumbnails) // [md-viewer](#markdown-viewer) // [search](#searching) // [fsearch](#file-search) // [zip-DL](#zip-downloads) // [ie4](#browser-support)
|
||||||
|
|
||||||
|
|
||||||
## readme toc
|
## readme toc
|
||||||
@@ -23,11 +23,13 @@ turn your phone or raspi into a portable file server with resumable uploads/down
|
|||||||
* [notes](#notes)
|
* [notes](#notes)
|
||||||
* [status](#status)
|
* [status](#status)
|
||||||
* [bugs](#bugs)
|
* [bugs](#bugs)
|
||||||
|
* [general bugs](#general-bugs)
|
||||||
* [not my bugs](#not-my-bugs)
|
* [not my bugs](#not-my-bugs)
|
||||||
* [the browser](#the-browser)
|
* [the browser](#the-browser)
|
||||||
* [tabs](#tabs)
|
* [tabs](#tabs)
|
||||||
* [hotkeys](#hotkeys)
|
* [hotkeys](#hotkeys)
|
||||||
* [tree-mode](#tree-mode)
|
* [tree-mode](#tree-mode)
|
||||||
|
* [thumbnails](#thumbnails)
|
||||||
* [zip downloads](#zip-downloads)
|
* [zip downloads](#zip-downloads)
|
||||||
* [uploading](#uploading)
|
* [uploading](#uploading)
|
||||||
* [file-search](#file-search)
|
* [file-search](#file-search)
|
||||||
@@ -35,6 +37,7 @@ turn your phone or raspi into a portable file server with resumable uploads/down
|
|||||||
* [other tricks](#other-tricks)
|
* [other tricks](#other-tricks)
|
||||||
* [searching](#searching)
|
* [searching](#searching)
|
||||||
* [search configuration](#search-configuration)
|
* [search configuration](#search-configuration)
|
||||||
|
* [database location](#database-location)
|
||||||
* [metadata from audio files](#metadata-from-audio-files)
|
* [metadata from audio files](#metadata-from-audio-files)
|
||||||
* [file parser plugins](#file-parser-plugins)
|
* [file parser plugins](#file-parser-plugins)
|
||||||
* [complete examples](#complete-examples)
|
* [complete examples](#complete-examples)
|
||||||
@@ -42,6 +45,8 @@ turn your phone or raspi into a portable file server with resumable uploads/down
|
|||||||
* [client examples](#client-examples)
|
* [client examples](#client-examples)
|
||||||
* [up2k](#up2k)
|
* [up2k](#up2k)
|
||||||
* [dependencies](#dependencies)
|
* [dependencies](#dependencies)
|
||||||
|
* [optional dependencies](#optional-dependencies)
|
||||||
|
* [install recommended deps](#install-recommended-deps)
|
||||||
* [optional gpl stuff](#optional-gpl-stuff)
|
* [optional gpl stuff](#optional-gpl-stuff)
|
||||||
* [sfx](#sfx)
|
* [sfx](#sfx)
|
||||||
* [sfx repack](#sfx-repack)
|
* [sfx repack](#sfx-repack)
|
||||||
@@ -55,25 +60,31 @@ turn your phone or raspi into a portable file server with resumable uploads/down
|
|||||||
|
|
||||||
download [copyparty-sfx.py](https://github.com/9001/copyparty/releases/latest/download/copyparty-sfx.py) and you're all set!
|
download [copyparty-sfx.py](https://github.com/9001/copyparty/releases/latest/download/copyparty-sfx.py) and you're all set!
|
||||||
|
|
||||||
running the sfx without arguments (for example doubleclicking it on Windows) will let anyone access the current folder; see `-h` for help if you want accounts and volumes etc
|
running the sfx without arguments (for example doubleclicking it on Windows) will give everyone full access to the current folder; see `-h` for help if you want accounts and volumes etc
|
||||||
|
|
||||||
you may also want these, especially on servers:
|
you may also want these, especially on servers:
|
||||||
* [contrib/systemd/copyparty.service](contrib/systemd/copyparty.service) to run copyparty as a systemd service
|
* [contrib/systemd/copyparty.service](contrib/systemd/copyparty.service) to run copyparty as a systemd service
|
||||||
* [contrib/nginx/copyparty.conf](contrib/nginx/copyparty.conf) to reverse-proxy behind nginx (for legit https)
|
* [contrib/nginx/copyparty.conf](contrib/nginx/copyparty.conf) to reverse-proxy behind nginx (for better https)
|
||||||
|
|
||||||
|
|
||||||
## notes
|
## notes
|
||||||
|
|
||||||
|
general:
|
||||||
|
* paper-printing is affected by dark/light-mode! use lightmode for color, darkmode for grayscale
|
||||||
|
* because no browsers currently implement the media-query to do this properly orz
|
||||||
|
|
||||||
|
browser-specific:
|
||||||
* iPhone/iPad: use Firefox to download files
|
* iPhone/iPad: use Firefox to download files
|
||||||
* Android-Chrome: increase "parallel uploads" for higher speed (android bug)
|
* Android-Chrome: increase "parallel uploads" for higher speed (android bug)
|
||||||
* Android-Firefox: takes a while to select files (their fix for ☝️)
|
* Android-Firefox: takes a while to select files (their fix for ☝️)
|
||||||
* Desktop-Firefox: ~~may use gigabytes of RAM if your files are massive~~ *seems to be OK now*
|
* Desktop-Firefox: ~~may use gigabytes of RAM if your files are massive~~ *seems to be OK now*
|
||||||
* paper-printing is affected by dark/light-mode! use lightmode for color, darkmode for grayscale
|
* Desktop-Firefox: may stop you from deleting folders you've uploaded until you visit `about:memory` and click `Minimize memory usage`
|
||||||
* because no browsers currently implement the media-query to do this properly orz
|
|
||||||
|
|
||||||
|
|
||||||
## status
|
## status
|
||||||
|
|
||||||
|
summary: all planned features work! now please enjoy the bloatening
|
||||||
|
|
||||||
* backend stuff
|
* backend stuff
|
||||||
* ☑ sanic multipart parser
|
* ☑ sanic multipart parser
|
||||||
* ☑ load balancer (multiprocessing)
|
* ☑ load balancer (multiprocessing)
|
||||||
@@ -91,9 +102,12 @@ you may also want these, especially on servers:
|
|||||||
* browser
|
* browser
|
||||||
* ☑ tree-view
|
* ☑ tree-view
|
||||||
* ☑ media player
|
* ☑ media player
|
||||||
* ✖ thumbnails
|
* ☑ thumbnails
|
||||||
* ✖ SPA (browse while uploading)
|
* ☑ images using Pillow
|
||||||
* currently safe using the file-tree on the left only, not folders in the file list
|
* ☑ videos using FFmpeg
|
||||||
|
* ☑ cache eviction (max-age; maybe max-size eventually)
|
||||||
|
* ☑ SPA (browse while uploading)
|
||||||
|
* if you use the file-tree on the left only, not folders in the file list
|
||||||
* server indexing
|
* server indexing
|
||||||
* ☑ locate files by contents
|
* ☑ locate files by contents
|
||||||
* ☑ search by name/path/date/size
|
* ☑ search by name/path/date/size
|
||||||
@@ -102,15 +116,18 @@ you may also want these, especially on servers:
|
|||||||
* ☑ viewer
|
* ☑ viewer
|
||||||
* ☑ editor (sure why not)
|
* ☑ editor (sure why not)
|
||||||
|
|
||||||
summary: it works! you can use it! (but technically not even close to beta)
|
|
||||||
|
|
||||||
|
|
||||||
# bugs
|
# bugs
|
||||||
|
|
||||||
* Windows: python 3.7 and older cannot read tags with ffprobe, so use mutagen or upgrade
|
* Windows: python 3.7 and older cannot read tags with ffprobe, so use mutagen or upgrade
|
||||||
* Windows: python 2.7 cannot index non-ascii filenames with `-e2d`
|
* Windows: python 2.7 cannot index non-ascii filenames with `-e2d`
|
||||||
* Windows: python 2.7 cannot handle filenames with mojibake
|
* Windows: python 2.7 cannot handle filenames with mojibake
|
||||||
* hiding the contents at url `/d1/d2/d3` using `-v :d1/d2/d3:cd2d` has the side-effect of creating databases (for files/tags) inside folders d1 and d2, and those databases take precedence over the main db at the top of the vfs - this means all files in d2 and below will be reindexed unless you already had a vfs entry at or below d2
|
* MacOS: `--th-ff-jpg` may fix thumbnails using macports-FFmpeg
|
||||||
|
|
||||||
|
## general bugs
|
||||||
|
|
||||||
|
* all volumes must exist / be available on startup; up2k (mtp especially) gets funky otherwise
|
||||||
|
* cannot mount something at `/d1/d2/d3` unless `d2` exists inside `d1`
|
||||||
* probably more, pls let me know
|
* probably more, pls let me know
|
||||||
|
|
||||||
## not my bugs
|
## not my bugs
|
||||||
@@ -139,11 +156,16 @@ summary: it works! you can use it! (but technically not even close to beta)
|
|||||||
the browser has the following hotkeys
|
the browser has the following hotkeys
|
||||||
* `I/K` prev/next folder
|
* `I/K` prev/next folder
|
||||||
* `P` parent folder
|
* `P` parent folder
|
||||||
|
* `G` toggle list / grid view
|
||||||
|
* `T` toggle thumbnails / icons
|
||||||
* when playing audio:
|
* when playing audio:
|
||||||
* `0..9` jump to 10%..90%
|
* `0..9` jump to 10%..90%
|
||||||
* `U/O` skip 10sec back/forward
|
* `U/O` skip 10sec back/forward
|
||||||
* `J/L` prev/next song
|
* `J/L` prev/next song
|
||||||
* `J` also starts playing the folder
|
* `J` also starts playing the folder
|
||||||
|
* in the grid view:
|
||||||
|
* `S` toggle multiselect
|
||||||
|
* `A/D` zoom
|
||||||
|
|
||||||
|
|
||||||
## tree-mode
|
## tree-mode
|
||||||
@@ -153,6 +175,15 @@ by default there's a breadcrumbs path; you can replace this with a tree-browser
|
|||||||
click `[-]` and `[+]` to adjust the size, and the `[a]` toggles if the tree should widen dynamically as you go deeper or stay fixed-size
|
click `[-]` and `[+]` to adjust the size, and the `[a]` toggles if the tree should widen dynamically as you go deeper or stay fixed-size
|
||||||
|
|
||||||
|
|
||||||
|
## thumbnails
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
it does static images with Pillow and uses FFmpeg for video files, so you may want to `--no-thumb` or maybe just `--no-vthumb` depending on how destructive your users are
|
||||||
|
|
||||||
|
images named `folder.jpg` and `folder.png` become the thumbnail of the folder they're in
|
||||||
|
|
||||||
|
|
||||||
## zip downloads
|
## zip downloads
|
||||||
|
|
||||||
the `zip` link next to folders can produce various types of zip/tar files using these alternatives in the browser settings tab:
|
the `zip` link next to folders can produce various types of zip/tar files using these alternatives in the browser settings tab:
|
||||||
@@ -176,8 +207,8 @@ you can also zip a selection of files or folders by clicking them in the browser
|
|||||||
## uploading
|
## uploading
|
||||||
|
|
||||||
two upload methods are available in the html client:
|
two upload methods are available in the html client:
|
||||||
* 🎈 bup, the basic uploader, supports almost every browser since netscape 4.0
|
* `🎈 bup`, the basic uploader, supports almost every browser since netscape 4.0
|
||||||
* 🚀 up2k, the fancy one
|
* `🚀 up2k`, the fancy one
|
||||||
|
|
||||||
up2k has several advantages:
|
up2k has several advantages:
|
||||||
* you can drop folders into the browser (files are added recursively)
|
* you can drop folders into the browser (files are added recursively)
|
||||||
@@ -212,14 +243,14 @@ and then theres the tabs below it,
|
|||||||
|
|
||||||

|

|
||||||
|
|
||||||
in the 🚀 up2k tab, after toggling the `[🔎]` switch green, any files/folders you drop onto the dropzone will be hashed on the client-side. Each hash is sent to the server which checks if that file exists somewhere already
|
in the `[🚀 up2k]` tab, after toggling the `[🔎]` switch green, any files/folders you drop onto the dropzone will be hashed on the client-side. Each hash is sent to the server which checks if that file exists somewhere already
|
||||||
|
|
||||||
files go into `[ok]` if they exist (and you get a link to where it is), otherwise they land in `[ng]`
|
files go into `[ok]` if they exist (and you get a link to where it is), otherwise they land in `[ng]`
|
||||||
* the main reason filesearch is combined with the uploader is cause the code was too spaghetti to separate it out somewhere else
|
* the main reason filesearch is combined with the uploader is cause the code was too spaghetti to separate it out somewhere else, this is no longer the case but now i've warmed up to the idea too much
|
||||||
|
|
||||||
adding the same file multiple times is blocked, so if you first search for a file and then decide to upload it, you have to click the `[cleanup]` button to discard `[done]` files
|
adding the same file multiple times is blocked, so if you first search for a file and then decide to upload it, you have to click the `[cleanup]` button to discard `[done]` files
|
||||||
|
|
||||||
note that since up2k has to read the file twice, 🎈 bup can be up to 2x faster if your internet connection is faster than the read-speed of your HDD
|
note that since up2k has to read the file twice, `[🎈 bup]` can be up to 2x faster in extreme cases (if your internet connection is faster than the read-speed of your HDD)
|
||||||
|
|
||||||
up2k has saved a few uploads from becoming corrupted in-transfer already; caught an android phone on wifi redhanded in wireshark with a bitflip, however bup with https would *probably* have noticed as well thanks to tls also functioning as an integrity check
|
up2k has saved a few uploads from becoming corrupted in-transfer already; caught an android phone on wifi redhanded in wireshark with a bitflip, however bup with https would *probably* have noticed as well thanks to tls also functioning as an integrity check
|
||||||
|
|
||||||
@@ -268,7 +299,29 @@ the same arguments can be set as volume flags, in addition to `d2d` and `d2t` fo
|
|||||||
* `-v ~/music::r:cd2d` disables **all** indexing, even if any `-e2*` are on
|
* `-v ~/music::r:cd2d` disables **all** indexing, even if any `-e2*` are on
|
||||||
* `-v ~/music::r:cd2t` disables all `-e2t*` (tags), does not affect `-e2d*`
|
* `-v ~/music::r:cd2t` disables all `-e2t*` (tags), does not affect `-e2d*`
|
||||||
|
|
||||||
`e2tsr` is probably always overkill, since `e2ds`/`e2dsa` would pick up any file modifications and cause `e2ts` to reindex those
|
note:
|
||||||
|
* `e2tsr` is probably always overkill, since `e2ds`/`e2dsa` would pick up any file modifications and cause `e2ts` to reindex those
|
||||||
|
* the rescan button in the admin panel has no effect unless the volume has `-e2ds` or higher
|
||||||
|
|
||||||
|
you can choose to only index filename/path/size/last-modified (and not the hash of the file contents) by setting `--no-hash` or the volume-flag `cdhash`, this has the following consequences:
|
||||||
|
* initial indexing is way faster, especially when the volume is on a networked disk
|
||||||
|
* makes it impossible to [file-search](#file-search)
|
||||||
|
* if someone uploads the same file contents, the upload will not be detected as a dupe, so it will not get symlinked or rejected
|
||||||
|
|
||||||
|
if you set `--no-hash`, you can enable hashing for specific volumes using flag `cehash`
|
||||||
|
|
||||||
|
|
||||||
|
## database location
|
||||||
|
|
||||||
|
copyparty creates a subfolder named `.hist` inside each volume where it stores the database, thumbnails, and some other stuff
|
||||||
|
|
||||||
|
this can instead be kept in a single place using the `--hist` argument, or the `hist=` volume flag, or a mix of both:
|
||||||
|
* `--hist ~/.cache/copyparty -v ~/music::r:chist=-` sets `~/.cache/copyparty` as the default place to put volume info, but `~/music` gets the regular `.hist` subfolder (`-` restores default behavior)
|
||||||
|
|
||||||
|
note:
|
||||||
|
* markdown edits are always stored in a local `.hist` subdirectory
|
||||||
|
* on windows the volflag path is cyglike, so `/c/temp` means `C:\temp` but use regular paths for `--hist`
|
||||||
|
* you can use cygpaths for volumes too, `-v C:\Users::r` and `-v /c/users::r` both work
|
||||||
|
|
||||||
|
|
||||||
## metadata from audio files
|
## metadata from audio files
|
||||||
@@ -287,6 +340,7 @@ see the beautiful mess of a dictionary in [mtag.py](https://github.com/9001/copy
|
|||||||
`--no-mutagen` disables mutagen and uses ffprobe instead, which...
|
`--no-mutagen` disables mutagen and uses ffprobe instead, which...
|
||||||
* is about 20x slower than mutagen
|
* is about 20x slower than mutagen
|
||||||
* catches a few tags that mutagen doesn't
|
* catches a few tags that mutagen doesn't
|
||||||
|
* melodic key, video resolution, framerate, pixfmt
|
||||||
* avoids pulling any GPL code into copyparty
|
* avoids pulling any GPL code into copyparty
|
||||||
* more importantly runs ffprobe on incoming files which is bad if your ffmpeg has a cve
|
* more importantly runs ffprobe on incoming files which is bad if your ffmpeg has a cve
|
||||||
|
|
||||||
@@ -299,9 +353,10 @@ copyparty can invoke external programs to collect additional metadata for files
|
|||||||
* `-mtp key=f,t5,~/bin/audio-key.py` uses `~/bin/audio-key.py` to get the `key` tag, replacing any existing metadata tag (`f,`), aborting if it takes longer than 5sec (`t5,`)
|
* `-mtp key=f,t5,~/bin/audio-key.py` uses `~/bin/audio-key.py` to get the `key` tag, replacing any existing metadata tag (`f,`), aborting if it takes longer than 5sec (`t5,`)
|
||||||
* `-v ~/music::r:cmtp=.bpm=~/bin/audio-bpm.py:cmtp=key=f,t5,~/bin/audio-key.py` both as a per-volume config wow this is getting ugly
|
* `-v ~/music::r:cmtp=.bpm=~/bin/audio-bpm.py:cmtp=key=f,t5,~/bin/audio-key.py` both as a per-volume config wow this is getting ugly
|
||||||
|
|
||||||
*but wait, there's more!* `-mtp` can be used for non-audio files as well using the `a` flag: `ay` only do audio files, `an` audio files are skipped, or `ad` always do it (d as in dontcare)
|
*but wait, there's more!* `-mtp` can be used for non-audio files as well using the `a` flag: `ay` only do audio files, `an` only do non-audio files, or `ad` do all files (d as in dontcare)
|
||||||
|
|
||||||
* `-mtp ext=an,~/bin/file-ext.py` runs `~/bin/file-ext.py` to get the `ext` tag only if file is not audio (`an`)
|
* `-mtp ext=an,~/bin/file-ext.py` runs `~/bin/file-ext.py` to get the `ext` tag only if file is not audio (`an`)
|
||||||
|
* `-mtp arch,built,ver,orig=an,eexe,edll,~/bin/exe.py` runs `~/bin/exe.py` to get properties about windows-binaries only if file is not audio (`an`) and file extension is exe or dll
|
||||||
|
|
||||||
|
|
||||||
## complete examples
|
## complete examples
|
||||||
@@ -338,14 +393,18 @@ copyparty can invoke external programs to collect additional metadata for files
|
|||||||
* `*2` using a wasm decoder which can sometimes get stuck and consumes a bit more power
|
* `*2` using a wasm decoder which can sometimes get stuck and consumes a bit more power
|
||||||
|
|
||||||
quick summary of more eccentric web-browsers trying to view a directory index:
|
quick summary of more eccentric web-browsers trying to view a directory index:
|
||||||
* safari (14.0.3/macos) is chrome with janky wasm, so playing opus can deadlock the javascript engine
|
|
||||||
* safari (14.0.1/iOS) same as macos, except it recovers from the deadlocks if you poke it a bit
|
| browser | will it blend |
|
||||||
* links (2.21/macports) can browse, login, upload/mkdir/msg
|
| ------- | ------------- |
|
||||||
* lynx (2.8.9/macports) can browse, login, upload/mkdir/msg
|
| **safari** (14.0.3/macos) | is chrome with janky wasm, so playing opus can deadlock the javascript engine |
|
||||||
* w3m (0.5.3/macports) can browse, login, upload at 100kB/s, mkdir/msg
|
| **safari** (14.0.1/iOS) | same as macos, except it recovers from the deadlocks if you poke it a bit |
|
||||||
* netsurf (3.10/arch) is basically ie6 with much better css (javascript has almost no effect)
|
| **links** (2.21/macports) | can browse, login, upload/mkdir/msg |
|
||||||
* ie4 and netscape 4.0 can browse (text is yellow on white), upload with `?b=u`
|
| **lynx** (2.8.9/macports) | can browse, login, upload/mkdir/msg |
|
||||||
* SerenityOS (22d13d8) hits a page fault, works with `?b=u`, file input not-impl, url params are multiplying
|
| **w3m** (0.5.3/macports) | can browse, login, upload at 100kB/s, mkdir/msg |
|
||||||
|
| **netsurf** (3.10/arch) | is basically ie6 with much better css (javascript has almost no effect) |
|
||||||
|
| **ie4** and **netscape** 4.0 | can browse (text is yellow on white), upload with `?b=u` |
|
||||||
|
| **SerenityOS** (22d13d8) | hits a page fault, works with `?b=u`, file input not-impl, url params are multiplying |
|
||||||
|
|
||||||
|
|
||||||
# client examples
|
# client examples
|
||||||
|
|
||||||
@@ -365,6 +424,8 @@ quick summary of more eccentric web-browsers trying to view a directory index:
|
|||||||
* cross-platform python client available in [./bin/](bin/)
|
* cross-platform python client available in [./bin/](bin/)
|
||||||
* [rclone](https://rclone.org/) as client can give ~5x performance, see [./docs/rclone.md](docs/rclone.md)
|
* [rclone](https://rclone.org/) as client can give ~5x performance, see [./docs/rclone.md](docs/rclone.md)
|
||||||
|
|
||||||
|
* sharex (screenshot utility): see [./contrib/sharex.sxcu](contrib/#sharexsxcu)
|
||||||
|
|
||||||
copyparty returns a truncated sha512sum of your PUT/POST as base64; you can generate the same checksum locally to verify uplaods:
|
copyparty returns a truncated sha512sum of your PUT/POST as base64; you can generate the same checksum locally to verify uplaods:
|
||||||
|
|
||||||
b512(){ printf "$((sha512sum||shasum -a512)|sed -E 's/ .*//;s/(..)/\\x\1/g')"|base64|head -c43;}
|
b512(){ printf "$((sha512sum||shasum -a512)|sed -E 's/ .*//;s/(..)/\\x\1/g')"|base64|head -c43;}
|
||||||
@@ -376,7 +437,7 @@ copyparty returns a truncated sha512sum of your PUT/POST as base64; you can gene
|
|||||||
quick outline of the up2k protocol, see [uploading](#uploading) for the web-client
|
quick outline of the up2k protocol, see [uploading](#uploading) for the web-client
|
||||||
* the up2k client splits a file into an "optimal" number of chunks
|
* the up2k client splits a file into an "optimal" number of chunks
|
||||||
* 1 MiB each, unless that becomes more than 256 chunks
|
* 1 MiB each, unless that becomes more than 256 chunks
|
||||||
* tries 1.5M, 2M, 3, 4, 6, ... until <= 256# or chunksize >= 32M
|
* tries 1.5M, 2M, 3, 4, 6, ... until <= 256 chunks or size >= 32M
|
||||||
* client posts the list of hashes, filename, size, last-modified
|
* client posts the list of hashes, filename, size, last-modified
|
||||||
* server creates the `wark`, an identifier for this upload
|
* server creates the `wark`, an identifier for this upload
|
||||||
* `sha512( salt + filesize + chunk_hashes )`
|
* `sha512( salt + filesize + chunk_hashes )`
|
||||||
@@ -391,13 +452,31 @@ quick outline of the up2k protocol, see [uploading](#uploading) for the web-clie
|
|||||||
|
|
||||||
* `jinja2` (is built into the SFX)
|
* `jinja2` (is built into the SFX)
|
||||||
|
|
||||||
**optional,** enables music tags:
|
|
||||||
|
## optional dependencies
|
||||||
|
|
||||||
|
enable music tags:
|
||||||
* either `mutagen` (fast, pure-python, skips a few tags, makes copyparty GPL? idk)
|
* either `mutagen` (fast, pure-python, skips a few tags, makes copyparty GPL? idk)
|
||||||
* or `FFprobe` (20x slower, more accurate, possibly dangerous depending on your distro and users)
|
* or `FFprobe` (20x slower, more accurate, possibly dangerous depending on your distro and users)
|
||||||
|
|
||||||
**optional,** will eventually enable thumbnails:
|
enable image thumbnails:
|
||||||
* `Pillow` (requires py2.7 or py3.5+)
|
* `Pillow` (requires py2.7 or py3.5+)
|
||||||
|
|
||||||
|
enable video thumbnails:
|
||||||
|
* `ffmpeg` and `ffprobe` somewhere in `$PATH`
|
||||||
|
|
||||||
|
enable reading HEIF pictures:
|
||||||
|
* `pyheif-pillow-opener` (requires Linux or a C compiler)
|
||||||
|
|
||||||
|
enable reading AVIF pictures:
|
||||||
|
* `pillow-avif-plugin`
|
||||||
|
|
||||||
|
|
||||||
|
## install recommended deps
|
||||||
|
```
|
||||||
|
python -m pip install --user -U jinja2 mutagen Pillow
|
||||||
|
```
|
||||||
|
|
||||||
|
|
||||||
## optional gpl stuff
|
## optional gpl stuff
|
||||||
|
|
||||||
@@ -409,8 +488,8 @@ these are standalone programs and will never be imported / evaluated by copypart
|
|||||||
# sfx
|
# sfx
|
||||||
|
|
||||||
currently there are two self-contained "binaries":
|
currently there are two self-contained "binaries":
|
||||||
* [copyparty-sfx.py](https://github.com/9001/copyparty/releases/latest/download/copyparty-sfx.py) -- pure python, works everywhere
|
* [copyparty-sfx.py](https://github.com/9001/copyparty/releases/latest/download/copyparty-sfx.py) -- pure python, works everywhere, **recommended**
|
||||||
* [copyparty-sfx.sh](https://github.com/9001/copyparty/releases/latest/download/copyparty-sfx.sh) -- smaller, but only for linux and macos
|
* [copyparty-sfx.sh](https://github.com/9001/copyparty/releases/latest/download/copyparty-sfx.sh) -- smaller, but only for linux and macos, kinda deprecated
|
||||||
|
|
||||||
launch either of them (**use sfx.py on systemd**) and it'll unpack and run copyparty, assuming you have python installed of course
|
launch either of them (**use sfx.py on systemd**) and it'll unpack and run copyparty, assuming you have python installed of course
|
||||||
|
|
||||||
@@ -470,20 +549,25 @@ in the `scripts` folder:
|
|||||||
|
|
||||||
roughly sorted by priority
|
roughly sorted by priority
|
||||||
|
|
||||||
* separate sqlite table per tag
|
|
||||||
* audio fingerprinting
|
|
||||||
* readme.md as epilogue
|
* readme.md as epilogue
|
||||||
|
* single sha512 across all up2k chunks? maybe
|
||||||
* reduce up2k roundtrips
|
* reduce up2k roundtrips
|
||||||
* start from a chunk index and just go
|
* start from a chunk index and just go
|
||||||
* terminate client on bad data
|
* terminate client on bad data
|
||||||
* `os.copy_file_range` for up2k cloning
|
|
||||||
* support pillow-simd
|
|
||||||
* figure out the deal with pixel3a not being connectable as hotspot
|
|
||||||
* pixel3a having unpredictable 3sec latency in general :||||
|
|
||||||
|
|
||||||
discarded ideas
|
discarded ideas
|
||||||
|
|
||||||
|
* separate sqlite table per tag
|
||||||
|
* performance fixed by skipping some indexes (`+mt.k`)
|
||||||
|
* audio fingerprinting
|
||||||
|
* only makes sense if there can be a wasm client and that doesn't exist yet (except for olaf which is agpl hence counts as not existing)
|
||||||
|
* `os.copy_file_range` for up2k cloning
|
||||||
|
* almost never hit this path anyways
|
||||||
* up2k partials ui
|
* up2k partials ui
|
||||||
|
* feels like there isn't much point
|
||||||
* cache sha512 chunks on client
|
* cache sha512 chunks on client
|
||||||
|
* too dangerous
|
||||||
* comment field
|
* comment field
|
||||||
|
* nah
|
||||||
* look into android thumbnail cache file format
|
* look into android thumbnail cache file format
|
||||||
|
* absolutely not
|
||||||
|
|||||||
@@ -45,3 +45,18 @@ you could replace winfsp with [dokan](https://github.com/dokan-dev/dokany/releas
|
|||||||
# [`mtag/`](mtag/)
|
# [`mtag/`](mtag/)
|
||||||
* standalone programs which perform misc. file analysis
|
* standalone programs which perform misc. file analysis
|
||||||
* copyparty can Popen programs like these during file indexing to collect additional metadata
|
* copyparty can Popen programs like these during file indexing to collect additional metadata
|
||||||
|
|
||||||
|
|
||||||
|
# [`dbtool.py`](dbtool.py)
|
||||||
|
upgrade utility which can show db info and help transfer data between databases, for example when a new version of copyparty recommends to wipe the DB and reindex because it now collects additional metadata during analysis, but you have some really expensive `-mtp` parsers and want to copy over the tags from the old db
|
||||||
|
|
||||||
|
for that example (upgrading to v0.11.0), first move the old db aside, launch copyparty, let it rebuild the db until the point where it starts running mtp (colored messages as it adds the mtp tags), then CTRL-C and patch in the old mtp tags from the old db instead
|
||||||
|
|
||||||
|
so assuming you have `-mtp` parsers to provide the tags `key` and `.bpm`:
|
||||||
|
|
||||||
|
```
|
||||||
|
~/bin/dbtool.py -ls up2k.db
|
||||||
|
~/bin/dbtool.py -src up2k.db.v0.10.22 up2k.db -cmp
|
||||||
|
~/bin/dbtool.py -src up2k.db.v0.10.22 up2k.db -rm-mtp-flag -copy key
|
||||||
|
~/bin/dbtool.py -src up2k.db.v0.10.22 up2k.db -rm-mtp-flag -copy .bpm -vac
|
||||||
|
```
|
||||||
|
|||||||
@@ -54,6 +54,12 @@ MACOS = platform.system() == "Darwin"
|
|||||||
info = log = dbg = None
|
info = log = dbg = None
|
||||||
|
|
||||||
|
|
||||||
|
print("{} v{} @ {}".format(
|
||||||
|
platform.python_implementation(),
|
||||||
|
".".join([str(x) for x in sys.version_info]),
|
||||||
|
sys.executable))
|
||||||
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from fuse import FUSE, FuseOSError, Operations
|
from fuse import FUSE, FuseOSError, Operations
|
||||||
except:
|
except:
|
||||||
|
|||||||
198
bin/dbtool.py
Executable file
198
bin/dbtool.py
Executable file
@@ -0,0 +1,198 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import sqlite3
|
||||||
|
import argparse
|
||||||
|
|
||||||
|
DB_VER = 3
|
||||||
|
|
||||||
|
|
||||||
|
def die(msg):
|
||||||
|
print("\033[31m\n" + msg + "\n\033[0m")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
|
||||||
|
def read_ver(db):
|
||||||
|
for tab in ["ki", "kv"]:
|
||||||
|
try:
|
||||||
|
c = db.execute(r"select v from {} where k = 'sver'".format(tab))
|
||||||
|
except:
|
||||||
|
continue
|
||||||
|
|
||||||
|
rows = c.fetchall()
|
||||||
|
if rows:
|
||||||
|
return int(rows[0][0])
|
||||||
|
|
||||||
|
return "corrupt"
|
||||||
|
|
||||||
|
|
||||||
|
def ls(db):
|
||||||
|
nfiles = next(db.execute("select count(w) from up"))[0]
|
||||||
|
ntags = next(db.execute("select count(w) from mt"))[0]
|
||||||
|
print(f"{nfiles} files")
|
||||||
|
print(f"{ntags} tags\n")
|
||||||
|
|
||||||
|
print("number of occurences for each tag,")
|
||||||
|
print(" 'x' = file has no tags")
|
||||||
|
print(" 't:mtp' = the mtp flag (file not mtp processed yet)")
|
||||||
|
print()
|
||||||
|
for k, nk in db.execute("select k, count(k) from mt group by k order by k"):
|
||||||
|
print(f"{nk:9} {k}")
|
||||||
|
|
||||||
|
|
||||||
|
def compare(n1, d1, n2, d2, verbose):
|
||||||
|
nt = next(d1.execute("select count(w) from up"))[0]
|
||||||
|
n = 0
|
||||||
|
miss = 0
|
||||||
|
for w, rd, fn in d1.execute("select w, rd, fn from up"):
|
||||||
|
n += 1
|
||||||
|
if n % 25_000 == 0:
|
||||||
|
m = f"\033[36mchecked {n:,} of {nt:,} files in {n1} against {n2}\033[0m"
|
||||||
|
print(m)
|
||||||
|
|
||||||
|
q = "select w from up where substr(w,1,16) = ?"
|
||||||
|
hit = d2.execute(q, (w[:16],)).fetchone()
|
||||||
|
if not hit:
|
||||||
|
miss += 1
|
||||||
|
if verbose:
|
||||||
|
print(f"file in {n1} missing in {n2}: [{w}] {rd}/{fn}")
|
||||||
|
|
||||||
|
print(f" {miss} files in {n1} missing in {n2}\n")
|
||||||
|
|
||||||
|
nt = next(d1.execute("select count(w) from mt"))[0]
|
||||||
|
n = 0
|
||||||
|
miss = {}
|
||||||
|
nmiss = 0
|
||||||
|
for w, k, v in d1.execute("select * from mt"):
|
||||||
|
n += 1
|
||||||
|
if n % 100_000 == 0:
|
||||||
|
m = f"\033[36mchecked {n:,} of {nt:,} tags in {n1} against {n2}, so far {nmiss} missing tags\033[0m"
|
||||||
|
print(m)
|
||||||
|
|
||||||
|
v2 = d2.execute("select v from mt where w = ? and +k = ?", (w, k)).fetchone()
|
||||||
|
if v2:
|
||||||
|
v2 = v2[0]
|
||||||
|
|
||||||
|
# if v != v2 and v2 and k in [".bpm", "key"] and n2 == "src":
|
||||||
|
# print(f"{w} [{rd}/{fn}] {k} = [{v}] / [{v2}]")
|
||||||
|
|
||||||
|
if v2 is not None:
|
||||||
|
if k.startswith("."):
|
||||||
|
try:
|
||||||
|
diff = abs(float(v) - float(v2))
|
||||||
|
if diff > float(v) / 0.9:
|
||||||
|
v2 = None
|
||||||
|
else:
|
||||||
|
v2 = v
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
if v != v2:
|
||||||
|
v2 = None
|
||||||
|
|
||||||
|
if v2 is None:
|
||||||
|
nmiss += 1
|
||||||
|
try:
|
||||||
|
miss[k] += 1
|
||||||
|
except:
|
||||||
|
miss[k] = 1
|
||||||
|
|
||||||
|
if verbose:
|
||||||
|
q = "select rd, fn from up where substr(w,1,16) = ?"
|
||||||
|
rd, fn = d1.execute(q, (w,)).fetchone()
|
||||||
|
print(f"missing in {n2}: [{w}] [{rd}/{fn}] {k} = {v}")
|
||||||
|
|
||||||
|
for k, v in sorted(miss.items()):
|
||||||
|
if v:
|
||||||
|
print(f"{n1} has {v:6} more {k:<6} tags than {n2}")
|
||||||
|
|
||||||
|
print(f"in total, {nmiss} missing tags in {n2}\n")
|
||||||
|
|
||||||
|
|
||||||
|
def copy_mtp(d1, d2, tag, rm):
|
||||||
|
nt = next(d1.execute("select count(w) from mt where k = ?", (tag,)))[0]
|
||||||
|
n = 0
|
||||||
|
ndone = 0
|
||||||
|
for w, k, v in d1.execute("select * from mt where k = ?", (tag,)):
|
||||||
|
n += 1
|
||||||
|
if n % 25_000 == 0:
|
||||||
|
m = f"\033[36m{n:,} of {nt:,} tags checked, so far {ndone} copied\033[0m"
|
||||||
|
print(m)
|
||||||
|
|
||||||
|
hit = d2.execute("select v from mt where w = ? and +k = ?", (w, k)).fetchone()
|
||||||
|
if hit:
|
||||||
|
hit = hit[0]
|
||||||
|
|
||||||
|
if hit != v:
|
||||||
|
ndone += 1
|
||||||
|
if hit is not None:
|
||||||
|
d2.execute("delete from mt where w = ? and +k = ?", (w, k))
|
||||||
|
|
||||||
|
d2.execute("insert into mt values (?,?,?)", (w, k, v))
|
||||||
|
if rm:
|
||||||
|
d2.execute("delete from mt where w = ? and +k = 't:mtp'", (w,))
|
||||||
|
|
||||||
|
d2.commit()
|
||||||
|
print(f"copied {ndone} {tag} tags over")
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
os.system("")
|
||||||
|
print()
|
||||||
|
|
||||||
|
ap = argparse.ArgumentParser()
|
||||||
|
ap.add_argument("db", help="database to work on")
|
||||||
|
ap.add_argument("-src", metavar="DB", type=str, help="database to copy from")
|
||||||
|
|
||||||
|
ap2 = ap.add_argument_group("informational / read-only stuff")
|
||||||
|
ap2.add_argument("-v", action="store_true", help="verbose")
|
||||||
|
ap2.add_argument("-ls", action="store_true", help="list summary for db")
|
||||||
|
ap2.add_argument("-cmp", action="store_true", help="compare databases")
|
||||||
|
|
||||||
|
ap2 = ap.add_argument_group("options which modify target db")
|
||||||
|
ap2.add_argument("-copy", metavar="TAG", type=str, help="mtp tag to copy over")
|
||||||
|
ap2.add_argument(
|
||||||
|
"-rm-mtp-flag",
|
||||||
|
action="store_true",
|
||||||
|
help="when an mtp tag is copied over, also mark that as done, so copyparty won't run mtp on it",
|
||||||
|
)
|
||||||
|
ap2.add_argument("-vac", action="store_true", help="optimize DB")
|
||||||
|
|
||||||
|
ar = ap.parse_args()
|
||||||
|
|
||||||
|
for v in [ar.db, ar.src]:
|
||||||
|
if v and not os.path.exists(v):
|
||||||
|
die("database must exist")
|
||||||
|
|
||||||
|
db = sqlite3.connect(ar.db)
|
||||||
|
ds = sqlite3.connect(ar.src) if ar.src else None
|
||||||
|
|
||||||
|
for d, n in [[ds, "src"], [db, "dst"]]:
|
||||||
|
if not d:
|
||||||
|
continue
|
||||||
|
|
||||||
|
ver = read_ver(d)
|
||||||
|
if ver == "corrupt":
|
||||||
|
die("{} database appears to be corrupt, sorry")
|
||||||
|
|
||||||
|
if ver != DB_VER:
|
||||||
|
m = f"{n} db is version {ver}, this tool only supports version {DB_VER}, please upgrade it with copyparty first"
|
||||||
|
die(m)
|
||||||
|
|
||||||
|
if ar.ls:
|
||||||
|
ls(db)
|
||||||
|
|
||||||
|
if ar.cmp:
|
||||||
|
if not ds:
|
||||||
|
die("need src db to compare against")
|
||||||
|
|
||||||
|
compare("src", ds, "dst", db, ar.v)
|
||||||
|
compare("dst", db, "src", ds, ar.v)
|
||||||
|
|
||||||
|
if ar.copy:
|
||||||
|
copy_mtp(ds, db, ar.copy, ar.rm_mtp_flag)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
96
bin/mtag/exe.py
Normal file
96
bin/mtag/exe.py
Normal file
@@ -0,0 +1,96 @@
|
|||||||
|
#!/usr/bin/env python
|
||||||
|
|
||||||
|
import sys
|
||||||
|
import time
|
||||||
|
import json
|
||||||
|
import pefile
|
||||||
|
|
||||||
|
"""
|
||||||
|
retrieve exe info,
|
||||||
|
example for multivalue providers
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
def unk(v):
|
||||||
|
return "unk({:04x})".format(v)
|
||||||
|
|
||||||
|
|
||||||
|
class PE2(pefile.PE):
|
||||||
|
def __init__(self, *a, **ka):
|
||||||
|
for k in [
|
||||||
|
# -- parse_data_directories:
|
||||||
|
"parse_import_directory",
|
||||||
|
"parse_export_directory",
|
||||||
|
# "parse_resources_directory",
|
||||||
|
"parse_debug_directory",
|
||||||
|
"parse_relocations_directory",
|
||||||
|
"parse_directory_tls",
|
||||||
|
"parse_directory_load_config",
|
||||||
|
"parse_delay_import_directory",
|
||||||
|
"parse_directory_bound_imports",
|
||||||
|
# -- full_load:
|
||||||
|
"parse_rich_header",
|
||||||
|
]:
|
||||||
|
setattr(self, k, self.noop)
|
||||||
|
|
||||||
|
super(PE2, self).__init__(*a, **ka)
|
||||||
|
|
||||||
|
def noop(*a, **ka):
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
try:
|
||||||
|
pe = PE2(sys.argv[1], fast_load=False)
|
||||||
|
except:
|
||||||
|
sys.exit(0)
|
||||||
|
|
||||||
|
arch = pe.FILE_HEADER.Machine
|
||||||
|
if arch == 0x14C:
|
||||||
|
arch = "x86"
|
||||||
|
elif arch == 0x8664:
|
||||||
|
arch = "x64"
|
||||||
|
else:
|
||||||
|
arch = unk(arch)
|
||||||
|
|
||||||
|
try:
|
||||||
|
buildtime = time.gmtime(pe.FILE_HEADER.TimeDateStamp)
|
||||||
|
buildtime = time.strftime("%Y-%m-%d_%H:%M:%S", buildtime)
|
||||||
|
except:
|
||||||
|
buildtime = "invalid"
|
||||||
|
|
||||||
|
ui = pe.OPTIONAL_HEADER.Subsystem
|
||||||
|
if ui == 2:
|
||||||
|
ui = "GUI"
|
||||||
|
elif ui == 3:
|
||||||
|
ui = "cmdline"
|
||||||
|
else:
|
||||||
|
ui = unk(ui)
|
||||||
|
|
||||||
|
extra = {}
|
||||||
|
if hasattr(pe, "FileInfo"):
|
||||||
|
for v1 in pe.FileInfo:
|
||||||
|
for v2 in v1:
|
||||||
|
if v2.name != "StringFileInfo":
|
||||||
|
continue
|
||||||
|
|
||||||
|
for v3 in v2.StringTable:
|
||||||
|
for k, v in v3.entries.items():
|
||||||
|
v = v.decode("utf-8", "replace").strip()
|
||||||
|
if not v:
|
||||||
|
continue
|
||||||
|
|
||||||
|
if k in [b"FileVersion", b"ProductVersion"]:
|
||||||
|
extra["ver"] = v
|
||||||
|
|
||||||
|
if k in [b"OriginalFilename", b"InternalName"]:
|
||||||
|
extra["orig"] = v
|
||||||
|
|
||||||
|
r = {
|
||||||
|
"arch": arch,
|
||||||
|
"built": buildtime,
|
||||||
|
"ui": ui,
|
||||||
|
"cksum": "{:08x}".format(pe.OPTIONAL_HEADER.CheckSum),
|
||||||
|
}
|
||||||
|
r.update(extra)
|
||||||
|
|
||||||
|
print(json.dumps(r, indent=4))
|
||||||
@@ -9,6 +9,16 @@
|
|||||||
* assumes the webserver and copyparty is running on the same server/IP
|
* assumes the webserver and copyparty is running on the same server/IP
|
||||||
* modify `10.13.1.1` as necessary if you wish to support browsers without javascript
|
* modify `10.13.1.1` as necessary if you wish to support browsers without javascript
|
||||||
|
|
||||||
|
### [`sharex.sxcu`](sharex.sxcu)
|
||||||
|
* sharex config file to upload screenshots and grab the URL
|
||||||
|
* `RequestURL`: full URL to the target folder
|
||||||
|
* `pw`: password (remove the `pw` line if anon-write)
|
||||||
|
|
||||||
|
however if your copyparty is behind a reverse-proxy, you may want to use [`sharex-html.sxcu`](sharex-html.sxcu) instead:
|
||||||
|
* `RequestURL`: full URL to the target folder
|
||||||
|
* `URL`: full URL to the root folder (with trailing slash) followed by `$regex:1|1$`
|
||||||
|
* `pw`: password (remove `Parameters` if anon-write)
|
||||||
|
|
||||||
### [`explorer-nothumbs-nofoldertypes.reg`](explorer-nothumbs-nofoldertypes.reg)
|
### [`explorer-nothumbs-nofoldertypes.reg`](explorer-nothumbs-nofoldertypes.reg)
|
||||||
* disables thumbnails and folder-type detection in windows explorer
|
* disables thumbnails and folder-type detection in windows explorer
|
||||||
* makes it way faster (especially for slow/networked locations (such as copyparty-fuse))
|
* makes it way faster (especially for slow/networked locations (such as copyparty-fuse))
|
||||||
|
|||||||
19
contrib/sharex-html.sxcu
Normal file
19
contrib/sharex-html.sxcu
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
{
|
||||||
|
"Version": "13.5.0",
|
||||||
|
"Name": "copyparty-html",
|
||||||
|
"DestinationType": "ImageUploader",
|
||||||
|
"RequestMethod": "POST",
|
||||||
|
"RequestURL": "http://127.0.0.1:3923/sharex",
|
||||||
|
"Parameters": {
|
||||||
|
"pw": "wark"
|
||||||
|
},
|
||||||
|
"Body": "MultipartFormData",
|
||||||
|
"Arguments": {
|
||||||
|
"act": "bput"
|
||||||
|
},
|
||||||
|
"FileFormName": "f",
|
||||||
|
"RegexList": [
|
||||||
|
"bytes // <a href=\"/([^\"]+)\""
|
||||||
|
],
|
||||||
|
"URL": "http://127.0.0.1:3923/$regex:1|1$"
|
||||||
|
}
|
||||||
17
contrib/sharex.sxcu
Normal file
17
contrib/sharex.sxcu
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
{
|
||||||
|
"Version": "13.5.0",
|
||||||
|
"Name": "copyparty",
|
||||||
|
"DestinationType": "ImageUploader",
|
||||||
|
"RequestMethod": "POST",
|
||||||
|
"RequestURL": "http://127.0.0.1:3923/sharex",
|
||||||
|
"Parameters": {
|
||||||
|
"pw": "wark",
|
||||||
|
"j": null
|
||||||
|
},
|
||||||
|
"Body": "MultipartFormData",
|
||||||
|
"Arguments": {
|
||||||
|
"act": "bput"
|
||||||
|
},
|
||||||
|
"FileFormName": "f",
|
||||||
|
"URL": "$json:files[0].url$"
|
||||||
|
}
|
||||||
@@ -2,6 +2,7 @@
|
|||||||
from __future__ import print_function, unicode_literals
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
import platform
|
import platform
|
||||||
|
import time
|
||||||
import sys
|
import sys
|
||||||
import os
|
import os
|
||||||
|
|
||||||
@@ -23,6 +24,7 @@ MACOS = platform.system() == "Darwin"
|
|||||||
|
|
||||||
class EnvParams(object):
|
class EnvParams(object):
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
|
self.t0 = time.time()
|
||||||
self.mod = os.path.dirname(os.path.realpath(__file__))
|
self.mod = os.path.dirname(os.path.realpath(__file__))
|
||||||
if self.mod.endswith("__init__"):
|
if self.mod.endswith("__init__"):
|
||||||
self.mod = os.path.dirname(self.mod)
|
self.mod = os.path.dirname(self.mod)
|
||||||
|
|||||||
@@ -225,6 +225,20 @@ def run_argparse(argv, formatter):
|
|||||||
--ciphers help = available ssl/tls ciphers,
|
--ciphers help = available ssl/tls ciphers,
|
||||||
--ssl-ver help = available ssl/tls versions,
|
--ssl-ver help = available ssl/tls versions,
|
||||||
default is what python considers safe, usually >= TLS1
|
default is what python considers safe, usually >= TLS1
|
||||||
|
|
||||||
|
values for --ls:
|
||||||
|
"USR" is a user to browse as; * is anonymous, ** is all users
|
||||||
|
"VOL" is a single volume to scan, default is * (all vols)
|
||||||
|
"FLAG" is flags;
|
||||||
|
"v" in addition to realpaths, print usernames and vpaths
|
||||||
|
"ln" only prints symlinks leaving the volume mountpoint
|
||||||
|
"p" exits 1 if any such symlinks are found
|
||||||
|
"r" resumes startup after the listing
|
||||||
|
examples:
|
||||||
|
--ls '**' # list all files which are possible to read
|
||||||
|
--ls '**,*,ln' # check for dangerous symlinks
|
||||||
|
--ls '**,*,ln,p,r' # check, then start normally if safe
|
||||||
|
\033[0m
|
||||||
"""
|
"""
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
@@ -249,6 +263,25 @@ def run_argparse(argv, formatter):
|
|||||||
ap.add_argument("--urlform", metavar="MODE", type=str, default="print,get", help="how to handle url-forms")
|
ap.add_argument("--urlform", metavar="MODE", type=str, default="print,get", help="how to handle url-forms")
|
||||||
ap.add_argument("--salt", type=str, default="hunter2", help="up2k file-hash salt")
|
ap.add_argument("--salt", type=str, default="hunter2", help="up2k file-hash salt")
|
||||||
|
|
||||||
|
ap2 = ap.add_argument_group('appearance options')
|
||||||
|
ap2.add_argument("--css-browser", metavar="L", help="URL to additional CSS to include")
|
||||||
|
|
||||||
|
ap2 = ap.add_argument_group('admin panel options')
|
||||||
|
ap2.add_argument("--no-rescan", action="store_true", help="disable ?scan (volume reindexing)")
|
||||||
|
ap2.add_argument("--no-stack", action="store_true", help="disable ?stack (list all stacks)")
|
||||||
|
|
||||||
|
ap2 = ap.add_argument_group('thumbnail options')
|
||||||
|
ap2.add_argument("--no-thumb", action="store_true", help="disable all thumbnails")
|
||||||
|
ap2.add_argument("--no-vthumb", action="store_true", help="disable video thumbnails")
|
||||||
|
ap2.add_argument("--th-size", metavar="WxH", default="320x256", help="thumbnail res")
|
||||||
|
ap2.add_argument("--th-no-crop", action="store_true", help="dynamic height; show full image")
|
||||||
|
ap2.add_argument("--th-no-jpg", action="store_true", help="disable jpg output")
|
||||||
|
ap2.add_argument("--th-no-webp", action="store_true", help="disable webp output")
|
||||||
|
ap2.add_argument("--th-ff-jpg", action="store_true", help="force jpg for video thumbs")
|
||||||
|
ap2.add_argument("--th-poke", metavar="SEC", type=int, default=300, help="activity labeling cooldown")
|
||||||
|
ap2.add_argument("--th-clean", metavar="SEC", type=int, default=43200, help="cleanup interval")
|
||||||
|
ap2.add_argument("--th-maxage", metavar="SEC", type=int, default=604800, help="max folder age")
|
||||||
|
|
||||||
ap2 = ap.add_argument_group('database options')
|
ap2 = ap.add_argument_group('database options')
|
||||||
ap2.add_argument("-e2d", action="store_true", help="enable up2k database")
|
ap2.add_argument("-e2d", action="store_true", help="enable up2k database")
|
||||||
ap2.add_argument("-e2ds", action="store_true", help="enable up2k db-scanner, sets -e2d")
|
ap2.add_argument("-e2ds", action="store_true", help="enable up2k db-scanner, sets -e2d")
|
||||||
@@ -256,11 +289,13 @@ def run_argparse(argv, formatter):
|
|||||||
ap2.add_argument("-e2t", action="store_true", help="enable metadata indexing")
|
ap2.add_argument("-e2t", action="store_true", help="enable metadata indexing")
|
||||||
ap2.add_argument("-e2ts", action="store_true", help="enable metadata scanner, sets -e2t")
|
ap2.add_argument("-e2ts", action="store_true", help="enable metadata scanner, sets -e2t")
|
||||||
ap2.add_argument("-e2tsr", action="store_true", help="rescan all metadata, sets -e2ts")
|
ap2.add_argument("-e2tsr", action="store_true", help="rescan all metadata, sets -e2ts")
|
||||||
|
ap2.add_argument("--hist", metavar="PATH", type=str, help="where to store volume state")
|
||||||
|
ap2.add_argument("--no-hash", action="store_true", help="disable hashing during e2ds folder scans")
|
||||||
ap2.add_argument("--no-mutagen", action="store_true", help="use ffprobe for tags instead")
|
ap2.add_argument("--no-mutagen", action="store_true", help="use ffprobe for tags instead")
|
||||||
ap2.add_argument("--no-mtag-mt", action="store_true", help="disable tag-read parallelism")
|
ap2.add_argument("--no-mtag-mt", action="store_true", help="disable tag-read parallelism")
|
||||||
ap2.add_argument("-mtm", metavar="M=t,t,t", action="append", type=str, help="add/replace metadata mapping")
|
ap2.add_argument("-mtm", metavar="M=t,t,t", action="append", type=str, help="add/replace metadata mapping")
|
||||||
ap2.add_argument("-mte", metavar="M,M,M", type=str, help="tags to index/display (comma-sep.)",
|
ap2.add_argument("-mte", metavar="M,M,M", type=str, help="tags to index/display (comma-sep.)",
|
||||||
default="circle,album,.tn,artist,title,.bpm,key,.dur,.q")
|
default="circle,album,.tn,artist,title,.bpm,key,.dur,.q,.vq,.aq,ac,vc,res,.fps")
|
||||||
ap2.add_argument("-mtp", metavar="M=[f,]bin", action="append", type=str, help="read tag M using bin")
|
ap2.add_argument("-mtp", metavar="M=[f,]bin", action="append", type=str, help="read tag M using bin")
|
||||||
ap2.add_argument("--srch-time", metavar="SEC", type=int, default=30, help="search deadline")
|
ap2.add_argument("--srch-time", metavar="SEC", type=int, default=30, help="search deadline")
|
||||||
|
|
||||||
@@ -273,11 +308,13 @@ def run_argparse(argv, formatter):
|
|||||||
ap2.add_argument("--ssl-log", metavar="PATH", help="log master secrets")
|
ap2.add_argument("--ssl-log", metavar="PATH", help="log master secrets")
|
||||||
|
|
||||||
ap2 = ap.add_argument_group('debug options')
|
ap2 = ap.add_argument_group('debug options')
|
||||||
|
ap2.add_argument("--ls", metavar="U[,V[,F]]", help="scan all volumes")
|
||||||
ap2.add_argument("--log-conn", action="store_true", help="print tcp-server msgs")
|
ap2.add_argument("--log-conn", action="store_true", help="print tcp-server msgs")
|
||||||
ap2.add_argument("--no-sendfile", action="store_true", help="disable sendfile")
|
ap2.add_argument("--no-sendfile", action="store_true", help="disable sendfile")
|
||||||
ap2.add_argument("--no-scandir", action="store_true", help="disable scandir")
|
ap2.add_argument("--no-scandir", action="store_true", help="disable scandir")
|
||||||
|
ap2.add_argument("--no-fastboot", action="store_true", help="wait for up2k indexing")
|
||||||
ap2.add_argument("--ihead", metavar="HEADER", action='append', help="dump incoming header")
|
ap2.add_argument("--ihead", metavar="HEADER", action='append', help="dump incoming header")
|
||||||
ap2.add_argument("--lf-url", metavar="RE", type=str, default=r"^/\.cpr/", help="dont log URLs matching")
|
ap2.add_argument("--lf-url", metavar="RE", type=str, default=r"^/\.cpr/|\?th=[wj]$", help="dont log URLs matching")
|
||||||
|
|
||||||
return ap.parse_args(args=argv[1:])
|
return ap.parse_args(args=argv[1:])
|
||||||
# fmt: on
|
# fmt: on
|
||||||
@@ -347,6 +384,9 @@ def main(argv=None):
|
|||||||
+ " (if you crash with codec errors then that is why)"
|
+ " (if you crash with codec errors then that is why)"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if WINDOWS and sys.version_info < (3, 6):
|
||||||
|
al.no_scandir = True
|
||||||
|
|
||||||
# signal.signal(signal.SIGINT, sighandler)
|
# signal.signal(signal.SIGINT, sighandler)
|
||||||
|
|
||||||
SvcHub(al).run()
|
SvcHub(al).run()
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
|
|
||||||
VERSION = (0, 10, 20)
|
VERSION = (0, 11, 12)
|
||||||
CODENAME = "zip it"
|
CODENAME = "the grid"
|
||||||
BUILD_DT = (2021, 5, 16)
|
BUILD_DT = (2021, 6, 12)
|
||||||
|
|
||||||
S_VERSION = ".".join(map(str, VERSION))
|
S_VERSION = ".".join(map(str, VERSION))
|
||||||
S_BUILD_DT = "{0:04d}-{1:02d}-{2:02d}".format(*BUILD_DT)
|
S_BUILD_DT = "{0:04d}-{1:02d}-{2:02d}".format(*BUILD_DT)
|
||||||
|
|||||||
@@ -5,35 +5,49 @@ import re
|
|||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import stat
|
import stat
|
||||||
|
import base64
|
||||||
|
import hashlib
|
||||||
import threading
|
import threading
|
||||||
|
|
||||||
from .__init__ import PY2, WINDOWS
|
from .__init__ import WINDOWS
|
||||||
from .util import IMPLICATIONS, undot, Pebkac, fsdec, fsenc, statdir, nuprint
|
from .util import IMPLICATIONS, undot, Pebkac, fsdec, fsenc, statdir, nuprint
|
||||||
|
|
||||||
|
|
||||||
class VFS(object):
|
class VFS(object):
|
||||||
"""single level in the virtual fs"""
|
"""single level in the virtual fs"""
|
||||||
|
|
||||||
def __init__(self, realpath, vpath, uread=[], uwrite=[], flags={}):
|
def __init__(self, realpath, vpath, uread=[], uwrite=[], uadm=[], flags={}):
|
||||||
self.realpath = realpath # absolute path on host filesystem
|
self.realpath = realpath # absolute path on host filesystem
|
||||||
self.vpath = vpath # absolute path in the virtual filesystem
|
self.vpath = vpath # absolute path in the virtual filesystem
|
||||||
self.uread = uread # users who can read this
|
self.uread = uread # users who can read this
|
||||||
self.uwrite = uwrite # users who can write this
|
self.uwrite = uwrite # users who can write this
|
||||||
|
self.uadm = uadm # users who are regular admins
|
||||||
self.flags = flags # config switches
|
self.flags = flags # config switches
|
||||||
self.nodes = {} # child nodes
|
self.nodes = {} # child nodes
|
||||||
|
self.histtab = None # all realpath->histpath
|
||||||
|
self.dbv = None # closest full/non-jump parent
|
||||||
|
|
||||||
|
if realpath:
|
||||||
|
self.histpath = os.path.join(realpath, ".hist") # db / thumbcache
|
||||||
self.all_vols = {vpath: self} # flattened recursive
|
self.all_vols = {vpath: self} # flattened recursive
|
||||||
|
else:
|
||||||
|
self.histpath = None
|
||||||
|
self.all_vols = None
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return "VFS({})".format(
|
return "VFS({})".format(
|
||||||
", ".join(
|
", ".join(
|
||||||
"{}={!r}".format(k, self.__dict__[k])
|
"{}={!r}".format(k, self.__dict__[k])
|
||||||
for k in "realpath vpath uread uwrite flags".split()
|
for k in "realpath vpath uread uwrite uadm flags".split()
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
def _trk(self, vol):
|
def get_all_vols(self, outdict):
|
||||||
self.all_vols[vol.vpath] = vol
|
if self.realpath:
|
||||||
return vol
|
outdict[self.vpath] = self
|
||||||
|
|
||||||
|
for v in self.nodes.values():
|
||||||
|
v.get_all_vols(outdict)
|
||||||
|
|
||||||
def add(self, src, dst):
|
def add(self, src, dst):
|
||||||
"""get existing, or add new path to the vfs"""
|
"""get existing, or add new path to the vfs"""
|
||||||
@@ -45,18 +59,19 @@ class VFS(object):
|
|||||||
name, dst = dst.split("/", 1)
|
name, dst = dst.split("/", 1)
|
||||||
if name in self.nodes:
|
if name in self.nodes:
|
||||||
# exists; do not manipulate permissions
|
# exists; do not manipulate permissions
|
||||||
return self._trk(self.nodes[name].add(src, dst))
|
return self.nodes[name].add(src, dst)
|
||||||
|
|
||||||
vn = VFS(
|
vn = VFS(
|
||||||
"{}/{}".format(self.realpath, name),
|
os.path.join(self.realpath, name) if self.realpath else None,
|
||||||
"{}/{}".format(self.vpath, name).lstrip("/"),
|
"{}/{}".format(self.vpath, name).lstrip("/"),
|
||||||
self.uread,
|
self.uread,
|
||||||
self.uwrite,
|
self.uwrite,
|
||||||
self.flags,
|
self.uadm,
|
||||||
|
self._copy_flags(name),
|
||||||
)
|
)
|
||||||
self._trk(vn)
|
vn.dbv = self.dbv or self
|
||||||
self.nodes[name] = vn
|
self.nodes[name] = vn
|
||||||
return self._trk(vn.add(src, dst))
|
return vn.add(src, dst)
|
||||||
|
|
||||||
if dst in self.nodes:
|
if dst in self.nodes:
|
||||||
# leaf exists; return as-is
|
# leaf exists; return as-is
|
||||||
@@ -65,8 +80,26 @@ class VFS(object):
|
|||||||
# leaf does not exist; create and keep permissions blank
|
# leaf does not exist; create and keep permissions blank
|
||||||
vp = "{}/{}".format(self.vpath, dst).lstrip("/")
|
vp = "{}/{}".format(self.vpath, dst).lstrip("/")
|
||||||
vn = VFS(src, vp)
|
vn = VFS(src, vp)
|
||||||
|
vn.dbv = self.dbv or self
|
||||||
self.nodes[dst] = vn
|
self.nodes[dst] = vn
|
||||||
return self._trk(vn)
|
return vn
|
||||||
|
|
||||||
|
def _copy_flags(self, name):
|
||||||
|
flags = {k: v for k, v in self.flags.items()}
|
||||||
|
hist = flags.get("hist")
|
||||||
|
if hist and hist != "-":
|
||||||
|
flags["hist"] = "{}/{}".format(hist.rstrip("/"), name)
|
||||||
|
|
||||||
|
return flags
|
||||||
|
|
||||||
|
def bubble_flags(self):
|
||||||
|
if self.dbv:
|
||||||
|
for k, v in self.dbv.flags.items():
|
||||||
|
if k not in ["hist"]:
|
||||||
|
self.flags[k] = v
|
||||||
|
|
||||||
|
for v in self.nodes.values():
|
||||||
|
v.bubble_flags()
|
||||||
|
|
||||||
def _find(self, vpath):
|
def _find(self, vpath):
|
||||||
"""return [vfs,remainder]"""
|
"""return [vfs,remainder]"""
|
||||||
@@ -94,6 +127,7 @@ class VFS(object):
|
|||||||
]
|
]
|
||||||
|
|
||||||
def get(self, vpath, uname, will_read, will_write):
|
def get(self, vpath, uname, will_read, will_write):
|
||||||
|
# type: (str, str, bool, bool) -> tuple[VFS, str]
|
||||||
"""returns [vfsnode,fs_remainder] if user has the requested permissions"""
|
"""returns [vfsnode,fs_remainder] if user has the requested permissions"""
|
||||||
vn, rem = self._find(vpath)
|
vn, rem = self._find(vpath)
|
||||||
|
|
||||||
@@ -105,6 +139,15 @@ class VFS(object):
|
|||||||
|
|
||||||
return vn, rem
|
return vn, rem
|
||||||
|
|
||||||
|
def get_dbv(self, vrem):
|
||||||
|
dbv = self.dbv
|
||||||
|
if not dbv:
|
||||||
|
return self, vrem
|
||||||
|
|
||||||
|
vrem = [self.vpath[len(dbv.vpath) + 1 :], vrem]
|
||||||
|
vrem = "/".join([x for x in vrem if x])
|
||||||
|
return dbv, vrem
|
||||||
|
|
||||||
def canonical(self, rem):
|
def canonical(self, rem):
|
||||||
"""returns the canonical path (fully-resolved absolute fs path)"""
|
"""returns the canonical path (fully-resolved absolute fs path)"""
|
||||||
rp = self.realpath
|
rp = self.realpath
|
||||||
@@ -133,7 +176,8 @@ class VFS(object):
|
|||||||
#
|
#
|
||||||
return os.path.realpath(rp)
|
return os.path.realpath(rp)
|
||||||
|
|
||||||
def ls(self, rem, uname, scandir, lstat=False):
|
def ls(self, rem, uname, scandir, incl_wo=False, lstat=False):
|
||||||
|
# type: (str, str, bool, bool, bool) -> tuple[str, str, dict[str, VFS]]
|
||||||
"""return user-readable [fsdir,real,virt] items at vpath"""
|
"""return user-readable [fsdir,real,virt] items at vpath"""
|
||||||
virt_vis = {} # nodes readable by user
|
virt_vis = {} # nodes readable by user
|
||||||
abspath = self.canonical(rem)
|
abspath = self.canonical(rem)
|
||||||
@@ -141,12 +185,12 @@ class VFS(object):
|
|||||||
real.sort()
|
real.sort()
|
||||||
if not rem:
|
if not rem:
|
||||||
for name, vn2 in sorted(self.nodes.items()):
|
for name, vn2 in sorted(self.nodes.items()):
|
||||||
if (
|
ok = uname in vn2.uread or "*" in vn2.uread
|
||||||
uname in vn2.uread
|
|
||||||
or "*" in vn2.uread
|
if not ok and incl_wo:
|
||||||
or uname in vn2.uwrite
|
ok = uname in vn2.uwrite or "*" in vn2.uwrite
|
||||||
or "*" in vn2.uwrite
|
|
||||||
):
|
if ok:
|
||||||
virt_vis[name] = vn2
|
virt_vis[name] = vn2
|
||||||
|
|
||||||
# no vfs nodes in the list of real inodes
|
# no vfs nodes in the list of real inodes
|
||||||
@@ -154,13 +198,21 @@ class VFS(object):
|
|||||||
|
|
||||||
return [abspath, real, virt_vis]
|
return [abspath, real, virt_vis]
|
||||||
|
|
||||||
def walk(self, rel, rem, uname, dots, scandir, lstat=False):
|
def walk(self, rel, rem, seen, uname, dots, scandir, lstat):
|
||||||
"""
|
"""
|
||||||
recursively yields from ./rem;
|
recursively yields from ./rem;
|
||||||
rel is a unix-style user-defined vpath (not vfs-related)
|
rel is a unix-style user-defined vpath (not vfs-related)
|
||||||
"""
|
"""
|
||||||
|
|
||||||
fsroot, vfs_ls, vfs_virt = self.ls(rem, uname, scandir, lstat)
|
fsroot, vfs_ls, vfs_virt = self.ls(
|
||||||
|
rem, uname, scandir, incl_wo=False, lstat=lstat
|
||||||
|
)
|
||||||
|
|
||||||
|
if seen and not fsroot.startswith(seen[-1]) and fsroot in seen:
|
||||||
|
print("bailing from symlink loop,\n {}\n {}".format(seen[-1], fsroot))
|
||||||
|
return
|
||||||
|
|
||||||
|
seen = seen[:] + [fsroot]
|
||||||
rfiles = [x for x in vfs_ls if not stat.S_ISDIR(x[1].st_mode)]
|
rfiles = [x for x in vfs_ls if not stat.S_ISDIR(x[1].st_mode)]
|
||||||
rdirs = [x for x in vfs_ls if stat.S_ISDIR(x[1].st_mode)]
|
rdirs = [x for x in vfs_ls if stat.S_ISDIR(x[1].st_mode)]
|
||||||
|
|
||||||
@@ -175,7 +227,7 @@ class VFS(object):
|
|||||||
|
|
||||||
wrel = (rel + "/" + rdir).lstrip("/")
|
wrel = (rel + "/" + rdir).lstrip("/")
|
||||||
wrem = (rem + "/" + rdir).lstrip("/")
|
wrem = (rem + "/" + rdir).lstrip("/")
|
||||||
for x in self.walk(wrel, wrem, uname, scandir, lstat):
|
for x in self.walk(wrel, wrem, seen, uname, dots, scandir, lstat):
|
||||||
yield x
|
yield x
|
||||||
|
|
||||||
for n, vfs in sorted(vfs_virt.items()):
|
for n, vfs in sorted(vfs_virt.items()):
|
||||||
@@ -183,14 +235,16 @@ class VFS(object):
|
|||||||
continue
|
continue
|
||||||
|
|
||||||
wrel = (rel + "/" + n).lstrip("/")
|
wrel = (rel + "/" + n).lstrip("/")
|
||||||
for x in vfs.walk(wrel, "", uname, scandir, lstat):
|
for x in vfs.walk(wrel, "", seen, uname, dots, scandir, lstat):
|
||||||
yield x
|
yield x
|
||||||
|
|
||||||
def zipgen(self, vrem, flt, uname, dots, scandir):
|
def zipgen(self, vrem, flt, uname, dots, scandir):
|
||||||
if flt:
|
if flt:
|
||||||
flt = {k: True for k in flt}
|
flt = {k: True for k in flt}
|
||||||
|
|
||||||
for vpath, apath, files, rd, vd in self.walk("", vrem, uname, dots, scandir):
|
for vpath, apath, files, rd, vd in self.walk(
|
||||||
|
"", vrem, [], uname, dots, scandir, False
|
||||||
|
):
|
||||||
if flt:
|
if flt:
|
||||||
files = [x for x in files if x[0] in flt]
|
files = [x for x in files if x[0] in flt]
|
||||||
|
|
||||||
@@ -226,17 +280,19 @@ class VFS(object):
|
|||||||
for f in [{"vp": v, "ap": a, "st": n[1]} for v, a, n in files]:
|
for f in [{"vp": v, "ap": a, "st": n[1]} for v, a, n in files]:
|
||||||
yield f
|
yield f
|
||||||
|
|
||||||
def user_tree(self, uname, readable=False, writable=False):
|
def user_tree(self, uname, readable, writable, admin):
|
||||||
ret = []
|
is_readable = False
|
||||||
opt1 = readable and (uname in self.uread or "*" in self.uread)
|
if uname in self.uread or "*" in self.uread:
|
||||||
opt2 = writable and (uname in self.uwrite or "*" in self.uwrite)
|
readable.append(self.vpath)
|
||||||
if opt1 or opt2:
|
is_readable = True
|
||||||
ret.append(self.vpath)
|
|
||||||
|
if uname in self.uwrite or "*" in self.uwrite:
|
||||||
|
writable.append(self.vpath)
|
||||||
|
if is_readable:
|
||||||
|
admin.append(self.vpath)
|
||||||
|
|
||||||
for _, vn in sorted(self.nodes.items()):
|
for _, vn in sorted(self.nodes.items()):
|
||||||
ret.extend(vn.user_tree(uname, readable, writable))
|
vn.user_tree(uname, readable, writable, admin)
|
||||||
|
|
||||||
return ret
|
|
||||||
|
|
||||||
|
|
||||||
class AuthSrv(object):
|
class AuthSrv(object):
|
||||||
@@ -257,6 +313,7 @@ class AuthSrv(object):
|
|||||||
self.reload()
|
self.reload()
|
||||||
|
|
||||||
def log(self, msg, c=0):
|
def log(self, msg, c=0):
|
||||||
|
if self.log_func:
|
||||||
self.log_func("auth", msg, c)
|
self.log_func("auth", msg, c)
|
||||||
|
|
||||||
def laggy_iter(self, iterable):
|
def laggy_iter(self, iterable):
|
||||||
@@ -269,7 +326,7 @@ class AuthSrv(object):
|
|||||||
|
|
||||||
yield prev, True
|
yield prev, True
|
||||||
|
|
||||||
def _parse_config_file(self, fd, user, mread, mwrite, mflags, mount):
|
def _parse_config_file(self, fd, user, mread, mwrite, madm, mflags, mount):
|
||||||
vol_src = None
|
vol_src = None
|
||||||
vol_dst = None
|
vol_dst = None
|
||||||
self.line_ctr = 0
|
self.line_ctr = 0
|
||||||
@@ -301,6 +358,7 @@ class AuthSrv(object):
|
|||||||
mount[vol_dst] = vol_src
|
mount[vol_dst] = vol_src
|
||||||
mread[vol_dst] = []
|
mread[vol_dst] = []
|
||||||
mwrite[vol_dst] = []
|
mwrite[vol_dst] = []
|
||||||
|
madm[vol_dst] = []
|
||||||
mflags[vol_dst] = {}
|
mflags[vol_dst] = {}
|
||||||
continue
|
continue
|
||||||
|
|
||||||
@@ -311,10 +369,15 @@ class AuthSrv(object):
|
|||||||
uname = "*"
|
uname = "*"
|
||||||
|
|
||||||
self._read_vol_str(
|
self._read_vol_str(
|
||||||
lvl, uname, mread[vol_dst], mwrite[vol_dst], mflags[vol_dst]
|
lvl,
|
||||||
|
uname,
|
||||||
|
mread[vol_dst],
|
||||||
|
mwrite[vol_dst],
|
||||||
|
madm[vol_dst],
|
||||||
|
mflags[vol_dst],
|
||||||
)
|
)
|
||||||
|
|
||||||
def _read_vol_str(self, lvl, uname, mr, mw, mf):
|
def _read_vol_str(self, lvl, uname, mr, mw, ma, mf):
|
||||||
if lvl == "c":
|
if lvl == "c":
|
||||||
cval = True
|
cval = True
|
||||||
if "=" in uname:
|
if "=" in uname:
|
||||||
@@ -332,6 +395,9 @@ class AuthSrv(object):
|
|||||||
if lvl in "wa":
|
if lvl in "wa":
|
||||||
mw.append(uname)
|
mw.append(uname)
|
||||||
|
|
||||||
|
if lvl == "a":
|
||||||
|
ma.append(uname)
|
||||||
|
|
||||||
def _read_volflag(self, flags, name, value, is_list):
|
def _read_volflag(self, flags, name, value, is_list):
|
||||||
if name not in ["mtp"]:
|
if name not in ["mtp"]:
|
||||||
flags[name] = value
|
flags[name] = value
|
||||||
@@ -355,6 +421,7 @@ class AuthSrv(object):
|
|||||||
user = {} # username:password
|
user = {} # username:password
|
||||||
mread = {} # mountpoint:[username]
|
mread = {} # mountpoint:[username]
|
||||||
mwrite = {} # mountpoint:[username]
|
mwrite = {} # mountpoint:[username]
|
||||||
|
madm = {} # mountpoint:[username]
|
||||||
mflags = {} # mountpoint:[flag]
|
mflags = {} # mountpoint:[flag]
|
||||||
mount = {} # dst:src (mountpoint:realpath)
|
mount = {} # dst:src (mountpoint:realpath)
|
||||||
|
|
||||||
@@ -372,23 +439,31 @@ class AuthSrv(object):
|
|||||||
raise Exception("invalid -v argument: [{}]".format(v_str))
|
raise Exception("invalid -v argument: [{}]".format(v_str))
|
||||||
|
|
||||||
src, dst, perms = m.groups()
|
src, dst, perms = m.groups()
|
||||||
|
if WINDOWS and src.startswith("/"):
|
||||||
|
src = "{}:\\{}".format(src[1], src[3:])
|
||||||
|
|
||||||
# print("\n".join([src, dst, perms]))
|
# print("\n".join([src, dst, perms]))
|
||||||
src = fsdec(os.path.abspath(fsenc(src)))
|
src = fsdec(os.path.abspath(fsenc(src)))
|
||||||
dst = dst.strip("/")
|
dst = dst.strip("/")
|
||||||
mount[dst] = src
|
mount[dst] = src
|
||||||
mread[dst] = []
|
mread[dst] = []
|
||||||
mwrite[dst] = []
|
mwrite[dst] = []
|
||||||
|
madm[dst] = []
|
||||||
mflags[dst] = {}
|
mflags[dst] = {}
|
||||||
|
|
||||||
perms = perms.split(":")
|
perms = perms.split(":")
|
||||||
for (lvl, uname) in [[x[0], x[1:]] for x in perms]:
|
for (lvl, uname) in [[x[0], x[1:]] for x in perms]:
|
||||||
self._read_vol_str(lvl, uname, mread[dst], mwrite[dst], mflags[dst])
|
self._read_vol_str(
|
||||||
|
lvl, uname, mread[dst], mwrite[dst], madm[dst], mflags[dst]
|
||||||
|
)
|
||||||
|
|
||||||
if self.args.c:
|
if self.args.c:
|
||||||
for cfg_fn in self.args.c:
|
for cfg_fn in self.args.c:
|
||||||
with open(cfg_fn, "rb") as f:
|
with open(cfg_fn, "rb") as f:
|
||||||
try:
|
try:
|
||||||
self._parse_config_file(f, user, mread, mwrite, mflags, mount)
|
self._parse_config_file(
|
||||||
|
f, user, mread, mwrite, madm, mflags, mount
|
||||||
|
)
|
||||||
except:
|
except:
|
||||||
m = "\n\033[1;31m\nerror in config file {} on line {}:\n\033[0m"
|
m = "\n\033[1;31m\nerror in config file {} on line {}:\n\033[0m"
|
||||||
print(m.format(cfg_fn, self.line_ctr))
|
print(m.format(cfg_fn, self.line_ctr))
|
||||||
@@ -399,7 +474,7 @@ class AuthSrv(object):
|
|||||||
vfs = VFS(os.path.abspath("."), "", ["*"], ["*"])
|
vfs = VFS(os.path.abspath("."), "", ["*"], ["*"])
|
||||||
elif "" not in mount:
|
elif "" not in mount:
|
||||||
# there's volumes but no root; make root inaccessible
|
# there's volumes but no root; make root inaccessible
|
||||||
vfs = VFS(os.path.abspath("."), "")
|
vfs = VFS(None, "")
|
||||||
vfs.flags["d2d"] = True
|
vfs.flags["d2d"] = True
|
||||||
|
|
||||||
maxdepth = 0
|
maxdepth = 0
|
||||||
@@ -410,13 +485,20 @@ class AuthSrv(object):
|
|||||||
|
|
||||||
if dst == "":
|
if dst == "":
|
||||||
# rootfs was mapped; fully replaces the default CWD vfs
|
# rootfs was mapped; fully replaces the default CWD vfs
|
||||||
vfs = VFS(mount[dst], dst, mread[dst], mwrite[dst], mflags[dst])
|
vfs = VFS(
|
||||||
|
mount[dst], dst, mread[dst], mwrite[dst], madm[dst], mflags[dst]
|
||||||
|
)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
v = vfs.add(mount[dst], dst)
|
v = vfs.add(mount[dst], dst)
|
||||||
v.uread = mread[dst]
|
v.uread = mread[dst]
|
||||||
v.uwrite = mwrite[dst]
|
v.uwrite = mwrite[dst]
|
||||||
|
v.uadm = madm[dst]
|
||||||
v.flags = mflags[dst]
|
v.flags = mflags[dst]
|
||||||
|
v.dbv = None
|
||||||
|
|
||||||
|
vfs.all_vols = {}
|
||||||
|
vfs.get_all_vols(vfs.all_vols)
|
||||||
|
|
||||||
missing_users = {}
|
missing_users = {}
|
||||||
for d in [mread, mwrite]:
|
for d in [mread, mwrite]:
|
||||||
@@ -433,6 +515,69 @@ class AuthSrv(object):
|
|||||||
)
|
)
|
||||||
raise Exception("invalid config")
|
raise Exception("invalid config")
|
||||||
|
|
||||||
|
promote = []
|
||||||
|
demote = []
|
||||||
|
for vol in vfs.all_vols.values():
|
||||||
|
hid = hashlib.sha512(fsenc(vol.realpath)).digest()
|
||||||
|
hid = base64.b32encode(hid).decode("ascii").lower()
|
||||||
|
vflag = vol.flags.get("hist")
|
||||||
|
if vflag == "-":
|
||||||
|
pass
|
||||||
|
elif vflag:
|
||||||
|
if WINDOWS and vflag.startswith("/"):
|
||||||
|
vflag = "{}:\\{}".format(vflag[1], vflag[3:])
|
||||||
|
vol.histpath = vflag
|
||||||
|
elif self.args.hist:
|
||||||
|
for nch in range(len(hid)):
|
||||||
|
hpath = os.path.join(self.args.hist, hid[: nch + 1])
|
||||||
|
try:
|
||||||
|
os.makedirs(hpath)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
powner = os.path.join(hpath, "owner.txt")
|
||||||
|
try:
|
||||||
|
with open(powner, "rb") as f:
|
||||||
|
owner = f.read().rstrip()
|
||||||
|
except:
|
||||||
|
owner = None
|
||||||
|
|
||||||
|
me = fsenc(vol.realpath).rstrip()
|
||||||
|
if owner not in [None, me]:
|
||||||
|
continue
|
||||||
|
|
||||||
|
if owner is None:
|
||||||
|
with open(powner, "wb") as f:
|
||||||
|
f.write(me)
|
||||||
|
|
||||||
|
vol.histpath = hpath
|
||||||
|
break
|
||||||
|
|
||||||
|
vol.histpath = os.path.realpath(vol.histpath)
|
||||||
|
if vol.dbv:
|
||||||
|
if os.path.exists(os.path.join(vol.histpath, "up2k.db")):
|
||||||
|
promote.append(vol)
|
||||||
|
vol.dbv = None
|
||||||
|
else:
|
||||||
|
demote.append(vol)
|
||||||
|
|
||||||
|
# discard jump-vols
|
||||||
|
for v in demote:
|
||||||
|
vfs.all_vols.pop(v.vpath)
|
||||||
|
|
||||||
|
if promote:
|
||||||
|
msg = [
|
||||||
|
"\n the following jump-volumes were generated to assist the vfs.\n As they contain a database (probably from v0.11.11 or older),\n they are promoted to full volumes:"
|
||||||
|
]
|
||||||
|
for vol in promote:
|
||||||
|
msg.append(
|
||||||
|
" /{} ({}) ({})".format(vol.vpath, vol.realpath, vol.histpath)
|
||||||
|
)
|
||||||
|
|
||||||
|
self.log("\n\n".join(msg) + "\n", c=3)
|
||||||
|
|
||||||
|
vfs.histtab = {v.realpath: v.histpath for v in vfs.all_vols.values()}
|
||||||
|
|
||||||
all_mte = {}
|
all_mte = {}
|
||||||
errors = False
|
errors = False
|
||||||
for vol in vfs.all_vols.values():
|
for vol in vfs.all_vols.values():
|
||||||
@@ -442,6 +587,10 @@ class AuthSrv(object):
|
|||||||
if self.args.e2d or "e2ds" in vol.flags:
|
if self.args.e2d or "e2ds" in vol.flags:
|
||||||
vol.flags["e2d"] = True
|
vol.flags["e2d"] = True
|
||||||
|
|
||||||
|
if self.args.no_hash:
|
||||||
|
if "ehash" not in vol.flags:
|
||||||
|
vol.flags["dhash"] = True
|
||||||
|
|
||||||
for k in ["e2t", "e2ts", "e2tsr"]:
|
for k in ["e2t", "e2ts", "e2tsr"]:
|
||||||
if getattr(self.args, k):
|
if getattr(self.args, k):
|
||||||
vol.flags[k] = True
|
vol.flags[k] = True
|
||||||
@@ -475,8 +624,10 @@ class AuthSrv(object):
|
|||||||
# verify tags mentioned by -mt[mp] are used by -mte
|
# verify tags mentioned by -mt[mp] are used by -mte
|
||||||
local_mtp = {}
|
local_mtp = {}
|
||||||
local_only_mtp = {}
|
local_only_mtp = {}
|
||||||
for a in vol.flags.get("mtp", []) + vol.flags.get("mtm", []):
|
tags = vol.flags.get("mtp", []) + vol.flags.get("mtm", [])
|
||||||
a = a.split("=")[0]
|
tags = [x.split("=")[0] for x in tags]
|
||||||
|
tags = [y for x in tags for y in x.split(",")]
|
||||||
|
for a in tags:
|
||||||
local_mtp[a] = True
|
local_mtp[a] = True
|
||||||
local = True
|
local = True
|
||||||
for b in self.args.mtp or []:
|
for b in self.args.mtp or []:
|
||||||
@@ -505,8 +656,10 @@ class AuthSrv(object):
|
|||||||
self.log(m.format(vol.vpath, mtp), 1)
|
self.log(m.format(vol.vpath, mtp), 1)
|
||||||
errors = True
|
errors = True
|
||||||
|
|
||||||
for mtp in self.args.mtp or []:
|
tags = self.args.mtp or []
|
||||||
mtp = mtp.split("=")[0]
|
tags = [x.split("=")[0] for x in tags]
|
||||||
|
tags = [y for x in tags for y in x.split(",")]
|
||||||
|
for mtp in tags:
|
||||||
if mtp not in all_mte:
|
if mtp not in all_mte:
|
||||||
m = 'metadata tag "{}" is defined by "-mtm" or "-mtp", but is not used by "-mte" (or by any "cmte" volume-flag)'
|
m = 'metadata tag "{}" is defined by "-mtm" or "-mtp", but is not used by "-mte" (or by any "cmte" volume-flag)'
|
||||||
self.log(m.format(mtp), 1)
|
self.log(m.format(mtp), 1)
|
||||||
@@ -515,6 +668,8 @@ class AuthSrv(object):
|
|||||||
if errors:
|
if errors:
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
|
vfs.bubble_flags()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
v, _ = vfs.get("/", "*", False, True)
|
v, _ = vfs.get("/", "*", False, True)
|
||||||
if self.warn_anonwrite and os.getcwd() == v.realpath:
|
if self.warn_anonwrite and os.getcwd() == v.realpath:
|
||||||
@@ -531,3 +686,90 @@ class AuthSrv(object):
|
|||||||
|
|
||||||
# import pprint
|
# import pprint
|
||||||
# pprint.pprint({"usr": user, "rd": mread, "wr": mwrite, "mnt": mount})
|
# pprint.pprint({"usr": user, "rd": mread, "wr": mwrite, "mnt": mount})
|
||||||
|
|
||||||
|
def dbg_ls(self):
|
||||||
|
users = self.args.ls
|
||||||
|
vols = "*"
|
||||||
|
flags = []
|
||||||
|
|
||||||
|
try:
|
||||||
|
users, vols = users.split(",", 1)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
try:
|
||||||
|
vols, flags = vols.split(",", 1)
|
||||||
|
flags = flags.split(",")
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
if users == "**":
|
||||||
|
users = list(self.user.keys()) + ["*"]
|
||||||
|
else:
|
||||||
|
users = [users]
|
||||||
|
|
||||||
|
for u in users:
|
||||||
|
if u not in self.user and u != "*":
|
||||||
|
raise Exception("user not found: " + u)
|
||||||
|
|
||||||
|
if vols == "*":
|
||||||
|
vols = ["/" + x for x in self.vfs.all_vols.keys()]
|
||||||
|
else:
|
||||||
|
vols = [vols]
|
||||||
|
|
||||||
|
for v in vols:
|
||||||
|
if not v.startswith("/"):
|
||||||
|
raise Exception("volumes must start with /")
|
||||||
|
|
||||||
|
if v[1:] not in self.vfs.all_vols:
|
||||||
|
raise Exception("volume not found: " + v)
|
||||||
|
|
||||||
|
self.log({"users": users, "vols": vols, "flags": flags})
|
||||||
|
for k, v in self.vfs.all_vols.items():
|
||||||
|
self.log("/{}: read({}) write({})".format(k, v.uread, v.uwrite))
|
||||||
|
|
||||||
|
flag_v = "v" in flags
|
||||||
|
flag_ln = "ln" in flags
|
||||||
|
flag_p = "p" in flags
|
||||||
|
flag_r = "r" in flags
|
||||||
|
|
||||||
|
n_bads = 0
|
||||||
|
for v in vols:
|
||||||
|
v = v[1:]
|
||||||
|
vtop = "/{}/".format(v) if v else "/"
|
||||||
|
for u in users:
|
||||||
|
self.log("checking /{} as {}".format(v, u))
|
||||||
|
try:
|
||||||
|
vn, _ = self.vfs.get(v, u, True, False)
|
||||||
|
except:
|
||||||
|
continue
|
||||||
|
|
||||||
|
atop = vn.realpath
|
||||||
|
g = vn.walk("", "", [], u, True, not self.args.no_scandir, False)
|
||||||
|
for vpath, apath, files, _, _ in g:
|
||||||
|
fnames = [n[0] for n in files]
|
||||||
|
vpaths = [vpath + "/" + n for n in fnames] if vpath else fnames
|
||||||
|
vpaths = [vtop + x for x in vpaths]
|
||||||
|
apaths = [os.path.join(apath, n) for n in fnames]
|
||||||
|
files = [[vpath + "/", apath + os.sep]] + list(zip(vpaths, apaths))
|
||||||
|
|
||||||
|
if flag_ln:
|
||||||
|
files = [x for x in files if not x[1].startswith(atop + os.sep)]
|
||||||
|
n_bads += len(files)
|
||||||
|
|
||||||
|
if flag_v:
|
||||||
|
msg = [
|
||||||
|
'# user "{}", vpath "{}"\n{}'.format(u, vp, ap)
|
||||||
|
for vp, ap in files
|
||||||
|
]
|
||||||
|
else:
|
||||||
|
msg = [x[1] for x in files]
|
||||||
|
|
||||||
|
if msg:
|
||||||
|
nuprint("\n".join(msg))
|
||||||
|
|
||||||
|
if n_bads and flag_p:
|
||||||
|
raise Exception("found symlink leaving volume, and strict is set")
|
||||||
|
|
||||||
|
if not flag_r:
|
||||||
|
sys.exit(0)
|
||||||
|
|||||||
@@ -44,7 +44,9 @@ class BrokerMp(object):
|
|||||||
proc.clients = {}
|
proc.clients = {}
|
||||||
proc.workload = 0
|
proc.workload = 0
|
||||||
|
|
||||||
thr = threading.Thread(target=self.collector, args=(proc,))
|
thr = threading.Thread(
|
||||||
|
target=self.collector, args=(proc,), name="mp-collector"
|
||||||
|
)
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
@@ -52,14 +54,19 @@ class BrokerMp(object):
|
|||||||
proc.start()
|
proc.start()
|
||||||
|
|
||||||
if not self.args.q:
|
if not self.args.q:
|
||||||
thr = threading.Thread(target=self.debug_load_balancer)
|
thr = threading.Thread(
|
||||||
|
target=self.debug_load_balancer, name="mp-dbg-loadbalancer"
|
||||||
|
)
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
def shutdown(self):
|
def shutdown(self):
|
||||||
self.log("broker", "shutting down")
|
self.log("broker", "shutting down")
|
||||||
for proc in self.procs:
|
for n, proc in enumerate(self.procs):
|
||||||
thr = threading.Thread(target=proc.q_pend.put([0, "shutdown", []]))
|
thr = threading.Thread(
|
||||||
|
target=proc.q_pend.put([0, "shutdown", []]),
|
||||||
|
name="mp-shutdown-{}-{}".format(n, len(self.procs)),
|
||||||
|
)
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
with self.mutex:
|
with self.mutex:
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
from __future__ import print_function, unicode_literals
|
from __future__ import print_function, unicode_literals
|
||||||
|
from copyparty.authsrv import AuthSrv
|
||||||
|
|
||||||
import sys
|
import sys
|
||||||
import time
|
import time
|
||||||
@@ -27,20 +28,23 @@ class MpWorker(object):
|
|||||||
self.retpend = {}
|
self.retpend = {}
|
||||||
self.retpend_mutex = threading.Lock()
|
self.retpend_mutex = threading.Lock()
|
||||||
self.mutex = threading.Lock()
|
self.mutex = threading.Lock()
|
||||||
self.workload_thr_active = False
|
self.workload_thr_alive = False
|
||||||
|
|
||||||
# we inherited signal_handler from parent,
|
# we inherited signal_handler from parent,
|
||||||
# replace it with something harmless
|
# replace it with something harmless
|
||||||
if not FAKE_MP:
|
if not FAKE_MP:
|
||||||
signal.signal(signal.SIGINT, self.signal_handler)
|
signal.signal(signal.SIGINT, self.signal_handler)
|
||||||
|
|
||||||
|
# starting to look like a good idea
|
||||||
|
self.asrv = AuthSrv(args, None, False)
|
||||||
|
|
||||||
# instantiate all services here (TODO: inheritance?)
|
# instantiate all services here (TODO: inheritance?)
|
||||||
self.httpsrv = HttpSrv(self)
|
self.httpsrv = HttpSrv(self, True)
|
||||||
self.httpsrv.disconnect_func = self.httpdrop
|
self.httpsrv.disconnect_func = self.httpdrop
|
||||||
|
|
||||||
# on winxp and some other platforms,
|
# on winxp and some other platforms,
|
||||||
# use thr.join() to block all signals
|
# use thr.join() to block all signals
|
||||||
thr = threading.Thread(target=self.main)
|
thr = threading.Thread(target=self.main, name="mpw-main")
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
thr.join()
|
thr.join()
|
||||||
@@ -79,9 +83,11 @@ class MpWorker(object):
|
|||||||
self.httpsrv.accept(sck, addr)
|
self.httpsrv.accept(sck, addr)
|
||||||
|
|
||||||
with self.mutex:
|
with self.mutex:
|
||||||
if not self.workload_thr_active:
|
if not self.workload_thr_alive:
|
||||||
self.workload_thr_alive = True
|
self.workload_thr_alive = True
|
||||||
thr = threading.Thread(target=self.thr_workload)
|
thr = threading.Thread(
|
||||||
|
target=self.thr_workload, name="mpw-workload"
|
||||||
|
)
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
|
|||||||
@@ -3,6 +3,7 @@ from __future__ import print_function, unicode_literals
|
|||||||
|
|
||||||
import threading
|
import threading
|
||||||
|
|
||||||
|
from .authsrv import AuthSrv
|
||||||
from .httpsrv import HttpSrv
|
from .httpsrv import HttpSrv
|
||||||
from .broker_util import ExceptionalQueue, try_exec
|
from .broker_util import ExceptionalQueue, try_exec
|
||||||
|
|
||||||
@@ -14,6 +15,7 @@ class BrokerThr(object):
|
|||||||
self.hub = hub
|
self.hub = hub
|
||||||
self.log = hub.log
|
self.log = hub.log
|
||||||
self.args = hub.args
|
self.args = hub.args
|
||||||
|
self.asrv = hub.asrv
|
||||||
|
|
||||||
self.mutex = threading.Lock()
|
self.mutex = threading.Lock()
|
||||||
|
|
||||||
|
|||||||
@@ -10,11 +10,13 @@ import json
|
|||||||
import string
|
import string
|
||||||
import socket
|
import socket
|
||||||
import ctypes
|
import ctypes
|
||||||
|
import traceback
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
import calendar
|
import calendar
|
||||||
|
|
||||||
from .__init__ import E, PY2, WINDOWS, ANYWIN
|
from .__init__ import E, PY2, WINDOWS, ANYWIN
|
||||||
from .util import * # noqa # pylint: disable=unused-wildcard-import
|
from .util import * # noqa # pylint: disable=unused-wildcard-import
|
||||||
|
from .authsrv import AuthSrv
|
||||||
from .szip import StreamZip
|
from .szip import StreamZip
|
||||||
from .star import StreamTar
|
from .star import StreamTar
|
||||||
|
|
||||||
@@ -22,6 +24,10 @@ if not PY2:
|
|||||||
unicode = str
|
unicode = str
|
||||||
|
|
||||||
|
|
||||||
|
NO_CACHE = {"Cache-Control": "no-cache"}
|
||||||
|
NO_STORE = {"Cache-Control": "no-store; max-age=0"}
|
||||||
|
|
||||||
|
|
||||||
class HttpCli(object):
|
class HttpCli(object):
|
||||||
"""
|
"""
|
||||||
Spawned by HttpConn to process one http transaction
|
Spawned by HttpConn to process one http transaction
|
||||||
@@ -30,12 +36,15 @@ class HttpCli(object):
|
|||||||
def __init__(self, conn):
|
def __init__(self, conn):
|
||||||
self.t0 = time.time()
|
self.t0 = time.time()
|
||||||
self.conn = conn
|
self.conn = conn
|
||||||
self.s = conn.s
|
self.s = conn.s # type: socket
|
||||||
self.sr = conn.sr
|
self.sr = conn.sr # type: Unrecv
|
||||||
self.ip = conn.addr[0]
|
self.ip = conn.addr[0]
|
||||||
self.addr = conn.addr
|
self.addr = conn.addr # type: tuple[str, int]
|
||||||
self.args = conn.args
|
self.args = conn.args
|
||||||
self.auth = conn.auth
|
self.is_mp = conn.is_mp
|
||||||
|
self.asrv = conn.asrv # type: AuthSrv
|
||||||
|
self.ico = conn.ico
|
||||||
|
self.thumbcli = conn.thumbcli
|
||||||
self.log_func = conn.log_func
|
self.log_func = conn.log_func
|
||||||
self.log_src = conn.log_src
|
self.log_src = conn.log_src
|
||||||
self.tls = hasattr(self.s, "cipher")
|
self.tls = hasattr(self.s, "cipher")
|
||||||
@@ -145,10 +154,9 @@ class HttpCli(object):
|
|||||||
self.vpath = unquotep(vpath)
|
self.vpath = unquotep(vpath)
|
||||||
|
|
||||||
pwd = uparam.get("pw")
|
pwd = uparam.get("pw")
|
||||||
self.uname = self.auth.iuser.get(pwd, "*")
|
self.uname = self.asrv.iuser.get(pwd, "*")
|
||||||
if self.uname:
|
self.rvol, self.wvol, self.avol = [[], [], []]
|
||||||
self.rvol = self.auth.vfs.user_tree(self.uname, readable=True)
|
self.asrv.vfs.user_tree(self.uname, self.rvol, self.wvol, self.avol)
|
||||||
self.wvol = self.auth.vfs.user_tree(self.uname, writable=True)
|
|
||||||
|
|
||||||
ua = self.headers.get("user-agent", "")
|
ua = self.headers.get("user-agent", "")
|
||||||
self.is_rclone = ua.startswith("rclone/")
|
self.is_rclone = ua.startswith("rclone/")
|
||||||
@@ -158,7 +166,7 @@ class HttpCli(object):
|
|||||||
uparam["b"] = False
|
uparam["b"] = False
|
||||||
cookies["b"] = False
|
cookies["b"] = False
|
||||||
|
|
||||||
self.do_log = not self.conn.lf_url or not self.conn.lf_url.match(self.req)
|
self.do_log = not self.conn.lf_url or not self.conn.lf_url.search(self.req)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if self.mode in ["GET", "HEAD"]:
|
if self.mode in ["GET", "HEAD"]:
|
||||||
@@ -250,7 +258,14 @@ class HttpCli(object):
|
|||||||
return "?" + "&".join(r)
|
return "?" + "&".join(r)
|
||||||
|
|
||||||
def redirect(
|
def redirect(
|
||||||
self, vpath, suf="", msg="aight", flavor="go to", click=True, use302=False
|
self,
|
||||||
|
vpath,
|
||||||
|
suf="",
|
||||||
|
msg="aight",
|
||||||
|
flavor="go to",
|
||||||
|
click=True,
|
||||||
|
status=200,
|
||||||
|
use302=False,
|
||||||
):
|
):
|
||||||
html = self.j2(
|
html = self.j2(
|
||||||
"msg",
|
"msg",
|
||||||
@@ -265,7 +280,7 @@ class HttpCli(object):
|
|||||||
h = {"Location": "/" + vpath, "Cache-Control": "no-cache"}
|
h = {"Location": "/" + vpath, "Cache-Control": "no-cache"}
|
||||||
self.reply(html, status=302, headers=h)
|
self.reply(html, status=302, headers=h)
|
||||||
else:
|
else:
|
||||||
self.reply(html)
|
self.reply(html, status=status)
|
||||||
|
|
||||||
def handle_get(self):
|
def handle_get(self):
|
||||||
if self.do_log:
|
if self.do_log:
|
||||||
@@ -283,6 +298,9 @@ class HttpCli(object):
|
|||||||
|
|
||||||
# "embedded" resources
|
# "embedded" resources
|
||||||
if self.vpath.startswith(".cpr"):
|
if self.vpath.startswith(".cpr"):
|
||||||
|
if self.vpath.startswith(".cpr/ico/"):
|
||||||
|
return self.tx_ico(self.vpath.split("/")[-1], exact=True)
|
||||||
|
|
||||||
static_path = os.path.join(E.mod, "web/", self.vpath[5:])
|
static_path = os.path.join(E.mod, "web/", self.vpath[5:])
|
||||||
return self.tx_file(static_path)
|
return self.tx_file(static_path)
|
||||||
|
|
||||||
@@ -303,9 +321,7 @@ class HttpCli(object):
|
|||||||
self.redirect(vpath, flavor="redirecting to", use302=True)
|
self.redirect(vpath, flavor="redirecting to", use302=True)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
self.readable, self.writable = self.conn.auth.vfs.can_access(
|
self.readable, self.writable = self.asrv.vfs.can_access(self.vpath, self.uname)
|
||||||
self.vpath, self.uname
|
|
||||||
)
|
|
||||||
if not self.readable and not self.writable:
|
if not self.readable and not self.writable:
|
||||||
if self.vpath:
|
if self.vpath:
|
||||||
self.log("inaccessible: [{}]".format(self.vpath))
|
self.log("inaccessible: [{}]".format(self.vpath))
|
||||||
@@ -317,6 +333,12 @@ class HttpCli(object):
|
|||||||
self.vpath = None
|
self.vpath = None
|
||||||
return self.tx_mounts()
|
return self.tx_mounts()
|
||||||
|
|
||||||
|
if "scan" in self.uparam:
|
||||||
|
return self.scanvol()
|
||||||
|
|
||||||
|
if "stack" in self.uparam:
|
||||||
|
return self.tx_stack()
|
||||||
|
|
||||||
return self.tx_browser()
|
return self.tx_browser()
|
||||||
|
|
||||||
def handle_options(self):
|
def handle_options(self):
|
||||||
@@ -416,18 +438,20 @@ class HttpCli(object):
|
|||||||
|
|
||||||
def dump_to_file(self):
|
def dump_to_file(self):
|
||||||
reader, remains = self.get_body_reader()
|
reader, remains = self.get_body_reader()
|
||||||
vfs, rem = self.conn.auth.vfs.get(self.vpath, self.uname, False, True)
|
vfs, rem = self.asrv.vfs.get(self.vpath, self.uname, False, True)
|
||||||
fdir = os.path.join(vfs.realpath, rem)
|
fdir = os.path.join(vfs.realpath, rem)
|
||||||
|
|
||||||
addr = self.ip.replace(":", ".")
|
addr = self.ip.replace(":", ".")
|
||||||
fn = "put-{:.6f}-{}.bin".format(time.time(), addr)
|
fn = "put-{:.6f}-{}.bin".format(time.time(), addr)
|
||||||
path = os.path.join(fdir, fn)
|
path = os.path.join(fdir, fn)
|
||||||
|
|
||||||
with open(path, "wb", 512 * 1024) as f:
|
with open(fsenc(path), "wb", 512 * 1024) as f:
|
||||||
post_sz, _, sha_b64 = hashcopy(self.conn, reader, f)
|
post_sz, _, sha_b64 = hashcopy(self.conn, reader, f)
|
||||||
|
|
||||||
|
vfs, vrem = vfs.get_dbv(rem)
|
||||||
|
|
||||||
self.conn.hsrv.broker.put(
|
self.conn.hsrv.broker.put(
|
||||||
False, "up2k.hash_file", vfs.realpath, vfs.flags, rem, fn
|
False, "up2k.hash_file", vfs.realpath, vfs.flags, vrem, fn
|
||||||
)
|
)
|
||||||
|
|
||||||
return post_sz, sha_b64, remains, path
|
return post_sz, sha_b64, remains, path
|
||||||
@@ -483,7 +507,7 @@ class HttpCli(object):
|
|||||||
if v is None:
|
if v is None:
|
||||||
raise Pebkac(422, "need zip or tar keyword")
|
raise Pebkac(422, "need zip or tar keyword")
|
||||||
|
|
||||||
vn, rem = self.auth.vfs.get(self.vpath, self.uname, True, False)
|
vn, rem = self.asrv.vfs.get(self.vpath, self.uname, True, False)
|
||||||
items = self.parser.require("files", 1024 * 1024)
|
items = self.parser.require("files", 1024 * 1024)
|
||||||
if not items:
|
if not items:
|
||||||
raise Pebkac(422, "need files list")
|
raise Pebkac(422, "need files list")
|
||||||
@@ -491,6 +515,7 @@ class HttpCli(object):
|
|||||||
items = items.replace("\r", "").split("\n")
|
items = items.replace("\r", "").split("\n")
|
||||||
items = [unquotep(x) for x in items if items]
|
items = [unquotep(x) for x in items if items]
|
||||||
|
|
||||||
|
self.parser.drop()
|
||||||
return self.tx_zip(k, v, vn, rem, items, self.args.ed)
|
return self.tx_zip(k, v, vn, rem, items, self.args.ed)
|
||||||
|
|
||||||
def handle_post_json(self):
|
def handle_post_json(self):
|
||||||
@@ -532,20 +557,21 @@ class HttpCli(object):
|
|||||||
self.vpath = "/".join([self.vpath, sub]).strip("/")
|
self.vpath = "/".join([self.vpath, sub]).strip("/")
|
||||||
body["name"] = name
|
body["name"] = name
|
||||||
|
|
||||||
vfs, rem = self.conn.auth.vfs.get(self.vpath, self.uname, False, True)
|
vfs, rem = self.asrv.vfs.get(self.vpath, self.uname, False, True)
|
||||||
|
dbv, vrem = vfs.get_dbv(rem)
|
||||||
|
|
||||||
body["vtop"] = vfs.vpath
|
body["vtop"] = dbv.vpath
|
||||||
body["ptop"] = vfs.realpath
|
body["ptop"] = dbv.realpath
|
||||||
body["prel"] = rem
|
body["prel"] = vrem
|
||||||
body["addr"] = self.ip
|
body["addr"] = self.ip
|
||||||
body["vcfg"] = vfs.flags
|
body["vcfg"] = dbv.flags
|
||||||
|
|
||||||
if sub:
|
if sub:
|
||||||
try:
|
try:
|
||||||
dst = os.path.join(vfs.realpath, rem)
|
dst = os.path.join(vfs.realpath, rem)
|
||||||
os.makedirs(dst)
|
os.makedirs(fsenc(dst))
|
||||||
except:
|
except:
|
||||||
if not os.path.isdir(dst):
|
if not os.path.isdir(fsenc(dst)):
|
||||||
raise Pebkac(400, "some file got your folder name")
|
raise Pebkac(400, "some file got your folder name")
|
||||||
|
|
||||||
x = self.conn.hsrv.broker.put(True, "up2k.handle_json", body)
|
x = self.conn.hsrv.broker.put(True, "up2k.handle_json", body)
|
||||||
@@ -560,8 +586,14 @@ class HttpCli(object):
|
|||||||
|
|
||||||
def handle_search(self, body):
|
def handle_search(self, body):
|
||||||
vols = []
|
vols = []
|
||||||
|
seen = {}
|
||||||
for vtop in self.rvol:
|
for vtop in self.rvol:
|
||||||
vfs, _ = self.conn.auth.vfs.get(vtop, self.uname, True, False)
|
vfs, _ = self.asrv.vfs.get(vtop, self.uname, True, False)
|
||||||
|
vfs = vfs.dbv or vfs
|
||||||
|
if vfs in seen:
|
||||||
|
continue
|
||||||
|
|
||||||
|
seen[vfs] = True
|
||||||
vols.append([vfs.vpath, vfs.realpath, vfs.flags])
|
vols.append([vfs.vpath, vfs.realpath, vfs.flags])
|
||||||
|
|
||||||
idx = self.conn.get_u2idx()
|
idx = self.conn.get_u2idx()
|
||||||
@@ -583,8 +615,9 @@ class HttpCli(object):
|
|||||||
taglist = {}
|
taglist = {}
|
||||||
else:
|
else:
|
||||||
# search by query params
|
# search by query params
|
||||||
self.log("qj: " + repr(body))
|
q = body["q"]
|
||||||
hits, taglist = idx.search(vols, body)
|
self.log("qj: " + q)
|
||||||
|
hits, taglist = idx.search(vols, q)
|
||||||
msg = len(hits)
|
msg = len(hits)
|
||||||
|
|
||||||
idx.p_end = time.time()
|
idx.p_end = time.time()
|
||||||
@@ -616,8 +649,8 @@ class HttpCli(object):
|
|||||||
except KeyError:
|
except KeyError:
|
||||||
raise Pebkac(400, "need hash and wark headers for binary POST")
|
raise Pebkac(400, "need hash and wark headers for binary POST")
|
||||||
|
|
||||||
vfs, _ = self.conn.auth.vfs.get(self.vpath, self.uname, False, True)
|
vfs, _ = self.asrv.vfs.get(self.vpath, self.uname, False, True)
|
||||||
ptop = vfs.realpath
|
ptop = (vfs.dbv or vfs).realpath
|
||||||
|
|
||||||
x = self.conn.hsrv.broker.put(True, "up2k.handle_chunk", ptop, wark, chash)
|
x = self.conn.hsrv.broker.put(True, "up2k.handle_chunk", ptop, wark, chash)
|
||||||
response = x.get()
|
response = x.get()
|
||||||
@@ -633,7 +666,7 @@ class HttpCli(object):
|
|||||||
|
|
||||||
reader = read_socket(self.sr, remains)
|
reader = read_socket(self.sr, remains)
|
||||||
|
|
||||||
with open(path, "rb+", 512 * 1024) as f:
|
with open(fsenc(path), "rb+", 512 * 1024) as f:
|
||||||
f.seek(cstart[0])
|
f.seek(cstart[0])
|
||||||
post_sz, _, sha_b64 = hashcopy(self.conn, reader, f)
|
post_sz, _, sha_b64 = hashcopy(self.conn, reader, f)
|
||||||
|
|
||||||
@@ -676,7 +709,7 @@ class HttpCli(object):
|
|||||||
times = (int(time.time()), int(lastmod))
|
times = (int(time.time()), int(lastmod))
|
||||||
self.log("no more chunks, setting times {}".format(times))
|
self.log("no more chunks, setting times {}".format(times))
|
||||||
try:
|
try:
|
||||||
os.utime(path, times)
|
os.utime(fsenc(path), times)
|
||||||
except:
|
except:
|
||||||
self.log("failed to utime ({}, {})".format(path, times))
|
self.log("failed to utime ({}, {})".format(path, times))
|
||||||
|
|
||||||
@@ -689,7 +722,7 @@ class HttpCli(object):
|
|||||||
pwd = self.parser.require("cppwd", 64)
|
pwd = self.parser.require("cppwd", 64)
|
||||||
self.parser.drop()
|
self.parser.drop()
|
||||||
|
|
||||||
if pwd in self.auth.iuser:
|
if pwd in self.asrv.iuser:
|
||||||
msg = "login ok"
|
msg = "login ok"
|
||||||
dt = datetime.utcfromtimestamp(time.time() + 60 * 60 * 24 * 365)
|
dt = datetime.utcfromtimestamp(time.time() + 60 * 60 * 24 * 365)
|
||||||
exp = dt.strftime("%a, %d %b %Y %H:%M:%S GMT")
|
exp = dt.strftime("%a, %d %b %Y %H:%M:%S GMT")
|
||||||
@@ -708,7 +741,7 @@ class HttpCli(object):
|
|||||||
self.parser.drop()
|
self.parser.drop()
|
||||||
|
|
||||||
nullwrite = self.args.nw
|
nullwrite = self.args.nw
|
||||||
vfs, rem = self.conn.auth.vfs.get(self.vpath, self.uname, False, True)
|
vfs, rem = self.asrv.vfs.get(self.vpath, self.uname, False, True)
|
||||||
self._assert_safe_rem(rem)
|
self._assert_safe_rem(rem)
|
||||||
|
|
||||||
sanitized = sanitize_fn(new_dir)
|
sanitized = sanitize_fn(new_dir)
|
||||||
@@ -737,7 +770,7 @@ class HttpCli(object):
|
|||||||
self.parser.drop()
|
self.parser.drop()
|
||||||
|
|
||||||
nullwrite = self.args.nw
|
nullwrite = self.args.nw
|
||||||
vfs, rem = self.conn.auth.vfs.get(self.vpath, self.uname, False, True)
|
vfs, rem = self.asrv.vfs.get(self.vpath, self.uname, False, True)
|
||||||
self._assert_safe_rem(rem)
|
self._assert_safe_rem(rem)
|
||||||
|
|
||||||
if not new_file.endswith(".md"):
|
if not new_file.endswith(".md"):
|
||||||
@@ -761,7 +794,7 @@ class HttpCli(object):
|
|||||||
|
|
||||||
def handle_plain_upload(self):
|
def handle_plain_upload(self):
|
||||||
nullwrite = self.args.nw
|
nullwrite = self.args.nw
|
||||||
vfs, rem = self.conn.auth.vfs.get(self.vpath, self.uname, False, True)
|
vfs, rem = self.asrv.vfs.get(self.vpath, self.uname, False, True)
|
||||||
self._assert_safe_rem(rem)
|
self._assert_safe_rem(rem)
|
||||||
|
|
||||||
files = []
|
files = []
|
||||||
@@ -798,8 +831,14 @@ class HttpCli(object):
|
|||||||
raise Pebkac(400, "empty files in post")
|
raise Pebkac(400, "empty files in post")
|
||||||
|
|
||||||
files.append([sz, sha512_hex, p_file, fname])
|
files.append([sz, sha512_hex, p_file, fname])
|
||||||
|
dbv, vrem = vfs.get_dbv(rem)
|
||||||
self.conn.hsrv.broker.put(
|
self.conn.hsrv.broker.put(
|
||||||
False, "up2k.hash_file", vfs.realpath, vfs.flags, rem, fname
|
False,
|
||||||
|
"up2k.hash_file",
|
||||||
|
dbv.realpath,
|
||||||
|
dbv.flags,
|
||||||
|
vrem,
|
||||||
|
fname,
|
||||||
)
|
)
|
||||||
self.conn.nbyte += sz
|
self.conn.nbyte += sz
|
||||||
|
|
||||||
@@ -829,18 +868,36 @@ class HttpCli(object):
|
|||||||
status = "OK"
|
status = "OK"
|
||||||
if errmsg:
|
if errmsg:
|
||||||
self.log(errmsg)
|
self.log(errmsg)
|
||||||
errmsg = "ERROR: " + errmsg
|
|
||||||
status = "ERROR"
|
status = "ERROR"
|
||||||
|
|
||||||
msg = "{} // {} bytes // {:.3f} MiB/s\n".format(status, sz_total, spd)
|
msg = "{} // {} bytes // {:.3f} MiB/s\n".format(status, sz_total, spd)
|
||||||
|
jmsg = {"status": status, "sz": sz_total, "mbps": round(spd, 3), "files": []}
|
||||||
|
|
||||||
|
if errmsg:
|
||||||
|
msg += errmsg + "\n"
|
||||||
|
jmsg["error"] = errmsg
|
||||||
|
errmsg = "ERROR: " + errmsg
|
||||||
|
|
||||||
for sz, sha512, ofn, lfn in files:
|
for sz, sha512, ofn, lfn in files:
|
||||||
vpath = self.vpath + "/" + lfn
|
vpath = (self.vpath + "/" if self.vpath else "") + lfn
|
||||||
msg += 'sha512: {} // {} bytes // <a href="/{}">{}</a>\n'.format(
|
msg += 'sha512: {} // {} bytes // <a href="/{}">{}</a>\n'.format(
|
||||||
sha512[:56], sz, quotep(vpath), html_escape(ofn, crlf=True)
|
sha512[:56], sz, quotep(vpath), html_escape(ofn, crlf=True)
|
||||||
)
|
)
|
||||||
# truncated SHA-512 prevents length extension attacks;
|
# truncated SHA-512 prevents length extension attacks;
|
||||||
# using SHA-512/224, optionally SHA-512/256 = :64
|
# using SHA-512/224, optionally SHA-512/256 = :64
|
||||||
|
jpart = {
|
||||||
|
"url": "{}://{}/{}".format(
|
||||||
|
"https" if self.tls else "http",
|
||||||
|
self.headers.get("host", "copyparty"),
|
||||||
|
vpath,
|
||||||
|
),
|
||||||
|
"sha512": sha512[:56],
|
||||||
|
"sz": sz,
|
||||||
|
"fn": lfn,
|
||||||
|
"fn_orig": ofn,
|
||||||
|
"path": vpath,
|
||||||
|
}
|
||||||
|
jmsg["files"].append(jpart)
|
||||||
|
|
||||||
vspd = self._spd(sz_total, False)
|
vspd = self._spd(sz_total, False)
|
||||||
self.log("{} {}".format(vspd, msg))
|
self.log("{} {}".format(vspd, msg))
|
||||||
@@ -852,7 +909,22 @@ class HttpCli(object):
|
|||||||
ft = "{}\n{}\n{}\n".format(ft, msg.rstrip(), errmsg)
|
ft = "{}\n{}\n{}\n".format(ft, msg.rstrip(), errmsg)
|
||||||
f.write(ft.encode("utf-8"))
|
f.write(ft.encode("utf-8"))
|
||||||
|
|
||||||
self.redirect(self.vpath, msg=msg, flavor="return to", click=False)
|
status = 400 if errmsg else 200
|
||||||
|
if "j" in self.uparam:
|
||||||
|
jtxt = json.dumps(jmsg, indent=2, sort_keys=True).encode("utf-8", "replace")
|
||||||
|
self.reply(jtxt, mime="application/json", status=status)
|
||||||
|
else:
|
||||||
|
self.redirect(
|
||||||
|
self.vpath,
|
||||||
|
msg=msg,
|
||||||
|
flavor="return to",
|
||||||
|
click=False,
|
||||||
|
status=status,
|
||||||
|
)
|
||||||
|
|
||||||
|
if errmsg:
|
||||||
|
return False
|
||||||
|
|
||||||
self.parser.drop()
|
self.parser.drop()
|
||||||
return True
|
return True
|
||||||
|
|
||||||
@@ -863,7 +935,7 @@ class HttpCli(object):
|
|||||||
raise Pebkac(400, "could not read lastmod from request")
|
raise Pebkac(400, "could not read lastmod from request")
|
||||||
|
|
||||||
nullwrite = self.args.nw
|
nullwrite = self.args.nw
|
||||||
vfs, rem = self.conn.auth.vfs.get(self.vpath, self.uname, False, True)
|
vfs, rem = self.asrv.vfs.get(self.vpath, self.uname, False, True)
|
||||||
self._assert_safe_rem(rem)
|
self._assert_safe_rem(rem)
|
||||||
|
|
||||||
# TODO:
|
# TODO:
|
||||||
@@ -927,16 +999,16 @@ class HttpCli(object):
|
|||||||
mdir, mfile = os.path.split(fp)
|
mdir, mfile = os.path.split(fp)
|
||||||
mfile2 = "{}.{:.3f}.md".format(mfile[:-3], srv_lastmod)
|
mfile2 = "{}.{:.3f}.md".format(mfile[:-3], srv_lastmod)
|
||||||
try:
|
try:
|
||||||
os.mkdir(os.path.join(mdir, ".hist"))
|
os.mkdir(fsenc(os.path.join(mdir, ".hist")))
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
os.rename(fp, os.path.join(mdir, ".hist", mfile2))
|
os.rename(fsenc(fp), fsenc(os.path.join(mdir, ".hist", mfile2)))
|
||||||
|
|
||||||
p_field, _, p_data = next(self.parser.gen)
|
p_field, _, p_data = next(self.parser.gen)
|
||||||
if p_field != "body":
|
if p_field != "body":
|
||||||
raise Pebkac(400, "expected body, got {}".format(p_field))
|
raise Pebkac(400, "expected body, got {}".format(p_field))
|
||||||
|
|
||||||
with open(fp, "wb", 512 * 1024) as f:
|
with open(fsenc(fp), "wb", 512 * 1024) as f:
|
||||||
sz, sha512, _ = hashcopy(self.conn, p_data, f)
|
sz, sha512, _ = hashcopy(self.conn, p_data, f)
|
||||||
|
|
||||||
new_lastmod = os.stat(fsenc(fp)).st_mtime
|
new_lastmod = os.stat(fsenc(fp)).st_mtime
|
||||||
@@ -952,14 +1024,13 @@ class HttpCli(object):
|
|||||||
return True
|
return True
|
||||||
|
|
||||||
def _chk_lastmod(self, file_ts):
|
def _chk_lastmod(self, file_ts):
|
||||||
date_fmt = "%a, %d %b %Y %H:%M:%S GMT"
|
file_lastmod = http_ts(file_ts)
|
||||||
file_dt = datetime.utcfromtimestamp(file_ts)
|
|
||||||
file_lastmod = file_dt.strftime(date_fmt)
|
|
||||||
|
|
||||||
cli_lastmod = self.headers.get("if-modified-since")
|
cli_lastmod = self.headers.get("if-modified-since")
|
||||||
if cli_lastmod:
|
if cli_lastmod:
|
||||||
try:
|
try:
|
||||||
cli_dt = time.strptime(cli_lastmod, date_fmt)
|
# some browser append "; length=573"
|
||||||
|
cli_lastmod = cli_lastmod.split(";")[0].strip()
|
||||||
|
cli_dt = time.strptime(cli_lastmod, HTTP_TS_FMT)
|
||||||
cli_ts = calendar.timegm(cli_dt)
|
cli_ts = calendar.timegm(cli_dt)
|
||||||
return file_lastmod, int(file_ts) > int(cli_ts)
|
return file_lastmod, int(file_ts) > int(cli_ts)
|
||||||
except Exception as ex:
|
except Exception as ex:
|
||||||
@@ -1106,13 +1177,13 @@ class HttpCli(object):
|
|||||||
# send reply
|
# send reply
|
||||||
|
|
||||||
if not is_compressed:
|
if not is_compressed:
|
||||||
self.out_headers["Cache-Control"] = "no-cache"
|
self.out_headers.update(NO_CACHE)
|
||||||
|
|
||||||
self.out_headers["Accept-Ranges"] = "bytes"
|
self.out_headers["Accept-Ranges"] = "bytes"
|
||||||
self.send_headers(
|
self.send_headers(
|
||||||
length=upper - lower,
|
length=upper - lower,
|
||||||
status=status,
|
status=status,
|
||||||
mime=guess_mime(req_path)[0] or "application/octet-stream",
|
mime=guess_mime(req_path),
|
||||||
)
|
)
|
||||||
|
|
||||||
logmsg += unicode(status) + logtail
|
logmsg += unicode(status) + logtail
|
||||||
@@ -1128,7 +1199,8 @@ class HttpCli(object):
|
|||||||
if use_sendfile:
|
if use_sendfile:
|
||||||
remains = sendfile_kern(lower, upper, f, self.s)
|
remains = sendfile_kern(lower, upper, f, self.s)
|
||||||
else:
|
else:
|
||||||
remains = sendfile_py(lower, upper, f, self.s)
|
actor = self.conn if self.is_mp else None
|
||||||
|
remains = sendfile_py(lower, upper, f, self.s, actor)
|
||||||
|
|
||||||
if remains > 0:
|
if remains > 0:
|
||||||
logmsg += " \033[31m" + unicode(upper - remains) + "\033[0m"
|
logmsg += " \033[31m" + unicode(upper - remains) + "\033[0m"
|
||||||
@@ -1202,6 +1274,34 @@ class HttpCli(object):
|
|||||||
self.log("{}, {}".format(logmsg, spd))
|
self.log("{}, {}".format(logmsg, spd))
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
def tx_ico(self, ext, exact=False):
|
||||||
|
if ext.endswith("/"):
|
||||||
|
ext = "folder"
|
||||||
|
exact = True
|
||||||
|
|
||||||
|
bad = re.compile(r"[](){}/[]|^[0-9_-]*$")
|
||||||
|
n = ext.split(".")[::-1]
|
||||||
|
if not exact:
|
||||||
|
n = n[:-1]
|
||||||
|
|
||||||
|
ext = ""
|
||||||
|
for v in n:
|
||||||
|
if len(v) > 7 or bad.search(v):
|
||||||
|
break
|
||||||
|
|
||||||
|
ext = "{}.{}".format(v, ext)
|
||||||
|
|
||||||
|
ext = ext.rstrip(".") or "unk"
|
||||||
|
if len(ext) > 11:
|
||||||
|
ext = "⋯" + ext[-9:]
|
||||||
|
|
||||||
|
mime, ico = self.ico.get(ext, not exact)
|
||||||
|
|
||||||
|
dt = datetime.utcfromtimestamp(E.t0)
|
||||||
|
lm = dt.strftime("%a, %d %b %Y %H:%M:%S GMT")
|
||||||
|
self.reply(ico, mime=mime, headers={"Last-Modified": lm})
|
||||||
|
return True
|
||||||
|
|
||||||
def tx_md(self, fs_path):
|
def tx_md(self, fs_path):
|
||||||
logmsg = "{:4} {} ".format("", self.req)
|
logmsg = "{:4} {} ".format("", self.req)
|
||||||
|
|
||||||
@@ -1224,7 +1324,7 @@ class HttpCli(object):
|
|||||||
file_ts = max(ts_md, ts_html)
|
file_ts = max(ts_md, ts_html)
|
||||||
file_lastmod, do_send = self._chk_lastmod(file_ts)
|
file_lastmod, do_send = self._chk_lastmod(file_ts)
|
||||||
self.out_headers["Last-Modified"] = file_lastmod
|
self.out_headers["Last-Modified"] = file_lastmod
|
||||||
self.out_headers["Cache-Control"] = "no-cache"
|
self.out_headers.update(NO_CACHE)
|
||||||
status = 200 if do_send else 304
|
status = 200 if do_send else 304
|
||||||
|
|
||||||
boundary = "\roll\tide"
|
boundary = "\roll\tide"
|
||||||
@@ -1236,7 +1336,7 @@ class HttpCli(object):
|
|||||||
"md_chk_rate": self.args.mcr,
|
"md_chk_rate": self.args.mcr,
|
||||||
"md": boundary,
|
"md": boundary,
|
||||||
}
|
}
|
||||||
html = template.render(**targs).encode("utf-8")
|
html = template.render(**targs).encode("utf-8", "replace")
|
||||||
html = html.split(boundary.encode("utf-8"))
|
html = html.split(boundary.encode("utf-8"))
|
||||||
if len(html) != 2:
|
if len(html) != 2:
|
||||||
raise Exception("boundary appears in " + html_path)
|
raise Exception("boundary appears in " + html_path)
|
||||||
@@ -1268,12 +1368,89 @@ class HttpCli(object):
|
|||||||
|
|
||||||
def tx_mounts(self):
|
def tx_mounts(self):
|
||||||
suf = self.urlq(rm=["h"])
|
suf = self.urlq(rm=["h"])
|
||||||
rvol = [x + "/" if x else x for x in self.rvol]
|
rvol, wvol, avol = [
|
||||||
wvol = [x + "/" if x else x for x in self.wvol]
|
[("/" + x).rstrip("/") + "/" for x in y]
|
||||||
html = self.j2("splash", this=self, rvol=rvol, wvol=wvol, url_suf=suf)
|
for y in [self.rvol, self.wvol, self.avol]
|
||||||
self.reply(html.encode("utf-8"))
|
]
|
||||||
|
|
||||||
|
if self.avol and not self.args.no_rescan:
|
||||||
|
x = self.conn.hsrv.broker.put(True, "up2k.get_state")
|
||||||
|
vs = json.loads(x.get())
|
||||||
|
vstate = {("/" + k).rstrip("/") + "/": v for k, v in vs["volstate"].items()}
|
||||||
|
else:
|
||||||
|
vstate = {}
|
||||||
|
vs = {"scanning": None, "hashq": None, "tagq": None, "mtpq": None}
|
||||||
|
|
||||||
|
html = self.j2(
|
||||||
|
"splash",
|
||||||
|
this=self,
|
||||||
|
rvol=rvol,
|
||||||
|
wvol=wvol,
|
||||||
|
avol=avol,
|
||||||
|
vstate=vstate,
|
||||||
|
scanning=vs["scanning"],
|
||||||
|
hashq=vs["hashq"],
|
||||||
|
tagq=vs["tagq"],
|
||||||
|
mtpq=vs["mtpq"],
|
||||||
|
url_suf=suf,
|
||||||
|
)
|
||||||
|
self.reply(html.encode("utf-8"), headers=NO_STORE)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
def scanvol(self):
|
||||||
|
if not self.readable or not self.writable:
|
||||||
|
raise Pebkac(403, "not admin")
|
||||||
|
|
||||||
|
if self.args.no_rescan:
|
||||||
|
raise Pebkac(403, "disabled by argv")
|
||||||
|
|
||||||
|
vn, _ = self.asrv.vfs.get(self.vpath, self.uname, True, True)
|
||||||
|
|
||||||
|
args = [self.asrv.vfs.all_vols, [vn.vpath]]
|
||||||
|
|
||||||
|
x = self.conn.hsrv.broker.put(True, "up2k.rescan", *args)
|
||||||
|
x = x.get()
|
||||||
|
if not x:
|
||||||
|
self.redirect("", "?h")
|
||||||
|
return ""
|
||||||
|
|
||||||
|
raise Pebkac(500, x)
|
||||||
|
|
||||||
|
def tx_stack(self):
|
||||||
|
if not self.readable or not self.writable:
|
||||||
|
raise Pebkac(403, "not admin")
|
||||||
|
|
||||||
|
if self.args.no_stack:
|
||||||
|
raise Pebkac(403, "disabled by argv")
|
||||||
|
|
||||||
|
threads = {}
|
||||||
|
names = dict([(t.ident, t.name) for t in threading.enumerate()])
|
||||||
|
for tid, stack in sys._current_frames().items():
|
||||||
|
name = "{} ({:x})".format(names.get(tid), tid)
|
||||||
|
threads[name] = stack
|
||||||
|
|
||||||
|
rret = []
|
||||||
|
bret = []
|
||||||
|
for name, stack in sorted(threads.items()):
|
||||||
|
ret = ["\n\n# {}".format(name)]
|
||||||
|
pad = None
|
||||||
|
for fn, lno, name, line in traceback.extract_stack(stack):
|
||||||
|
fn = os.sep.join(fn.split(os.sep)[-3:])
|
||||||
|
ret.append('File: "{}", line {}, in {}'.format(fn, lno, name))
|
||||||
|
if line:
|
||||||
|
ret.append(" " + str(line.strip()))
|
||||||
|
if "self.not_empty.wait()" in line:
|
||||||
|
pad = " " * 4
|
||||||
|
|
||||||
|
if pad:
|
||||||
|
bret += [ret[0]] + [pad + x for x in ret[1:]]
|
||||||
|
else:
|
||||||
|
rret += ret
|
||||||
|
|
||||||
|
ret = rret + bret
|
||||||
|
ret = ("<pre>" + "\n".join(ret)).encode("utf-8")
|
||||||
|
self.reply(ret)
|
||||||
|
|
||||||
def tx_tree(self):
|
def tx_tree(self):
|
||||||
top = self.uparam["tree"] or ""
|
top = self.uparam["tree"] or ""
|
||||||
dst = self.vpath
|
dst = self.vpath
|
||||||
@@ -1302,8 +1479,10 @@ class HttpCli(object):
|
|||||||
ret["k" + quotep(excl)] = sub
|
ret["k" + quotep(excl)] = sub
|
||||||
|
|
||||||
try:
|
try:
|
||||||
vn, rem = self.auth.vfs.get(top, self.uname, True, False)
|
vn, rem = self.asrv.vfs.get(top, self.uname, True, False)
|
||||||
fsroot, vfs_ls, vfs_virt = vn.ls(rem, self.uname, not self.args.no_scandir)
|
fsroot, vfs_ls, vfs_virt = vn.ls(
|
||||||
|
rem, self.uname, not self.args.no_scandir, incl_wo=True
|
||||||
|
)
|
||||||
except:
|
except:
|
||||||
vfs_ls = []
|
vfs_ls = []
|
||||||
vfs_virt = {}
|
vfs_virt = {}
|
||||||
@@ -1341,15 +1520,52 @@ class HttpCli(object):
|
|||||||
|
|
||||||
vpnodes.append([quotep(vpath) + "/", html_escape(node, crlf=True)])
|
vpnodes.append([quotep(vpath) + "/", html_escape(node, crlf=True)])
|
||||||
|
|
||||||
vn, rem = self.auth.vfs.get(
|
vn, rem = self.asrv.vfs.get(
|
||||||
self.vpath, self.uname, self.readable, self.writable
|
self.vpath, self.uname, self.readable, self.writable
|
||||||
)
|
)
|
||||||
abspath = vn.canonical(rem)
|
abspath = vn.canonical(rem)
|
||||||
|
dbv, vrem = vn.get_dbv(rem)
|
||||||
|
|
||||||
if not os.path.exists(fsenc(abspath)):
|
try:
|
||||||
# print(abspath)
|
st = os.stat(fsenc(abspath))
|
||||||
|
except:
|
||||||
raise Pebkac(404)
|
raise Pebkac(404)
|
||||||
|
|
||||||
|
if self.readable:
|
||||||
|
if rem.startswith(".hist/up2k."):
|
||||||
|
raise Pebkac(403)
|
||||||
|
|
||||||
|
is_dir = stat.S_ISDIR(st.st_mode)
|
||||||
|
th_fmt = self.uparam.get("th")
|
||||||
|
if th_fmt is not None:
|
||||||
|
if is_dir:
|
||||||
|
for fn in ["folder.png", "folder.jpg"]:
|
||||||
|
fp = os.path.join(abspath, fn)
|
||||||
|
if os.path.exists(fp):
|
||||||
|
vrem = "{}/{}".format(vrem.rstrip("/"), fn)
|
||||||
|
is_dir = False
|
||||||
|
break
|
||||||
|
|
||||||
|
if is_dir:
|
||||||
|
return self.tx_ico("a.folder")
|
||||||
|
|
||||||
|
thp = None
|
||||||
|
if self.thumbcli:
|
||||||
|
thp = self.thumbcli.get(
|
||||||
|
dbv.realpath, vrem, int(st.st_mtime), th_fmt
|
||||||
|
)
|
||||||
|
|
||||||
|
if thp:
|
||||||
|
return self.tx_file(thp)
|
||||||
|
|
||||||
|
return self.tx_ico(rem)
|
||||||
|
|
||||||
|
if not is_dir:
|
||||||
|
if abspath.endswith(".md") and "raw" not in self.uparam:
|
||||||
|
return self.tx_md(abspath)
|
||||||
|
|
||||||
|
return self.tx_file(abspath)
|
||||||
|
|
||||||
srv_info = []
|
srv_info = []
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@@ -1368,7 +1584,7 @@ class HttpCli(object):
|
|||||||
)
|
)
|
||||||
srv_info.append(humansize(bfree.value) + " free")
|
srv_info.append(humansize(bfree.value) + " free")
|
||||||
else:
|
else:
|
||||||
sv = os.statvfs(abspath)
|
sv = os.statvfs(fsenc(abspath))
|
||||||
free = humansize(sv.f_frsize * sv.f_bfree, True)
|
free = humansize(sv.f_frsize * sv.f_bfree, True)
|
||||||
total = humansize(sv.f_frsize * sv.f_blocks, True)
|
total = humansize(sv.f_frsize * sv.f_blocks, True)
|
||||||
|
|
||||||
@@ -1428,31 +1644,28 @@ class HttpCli(object):
|
|||||||
if not self.readable:
|
if not self.readable:
|
||||||
if is_ls:
|
if is_ls:
|
||||||
ret = json.dumps(ls_ret)
|
ret = json.dumps(ls_ret)
|
||||||
self.reply(ret.encode("utf-8", "replace"), mime="application/json")
|
self.reply(
|
||||||
|
ret.encode("utf-8", "replace"),
|
||||||
|
mime="application/json",
|
||||||
|
headers=NO_STORE,
|
||||||
|
)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
if not os.path.isdir(fsenc(abspath)):
|
if not stat.S_ISDIR(st.st_mode):
|
||||||
raise Pebkac(404)
|
raise Pebkac(404)
|
||||||
|
|
||||||
html = self.j2(tpl, **j2a)
|
html = self.j2(tpl, **j2a)
|
||||||
self.reply(html.encode("utf-8", "replace"))
|
self.reply(html.encode("utf-8", "replace"), headers=NO_STORE)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
if not os.path.isdir(fsenc(abspath)):
|
|
||||||
if abspath.endswith(".md") and "raw" not in self.uparam:
|
|
||||||
return self.tx_md(abspath)
|
|
||||||
|
|
||||||
if rem.startswith(".hist/up2k."):
|
|
||||||
raise Pebkac(403)
|
|
||||||
|
|
||||||
return self.tx_file(abspath)
|
|
||||||
|
|
||||||
for k in ["zip", "tar"]:
|
for k in ["zip", "tar"]:
|
||||||
v = self.uparam.get(k)
|
v = self.uparam.get(k)
|
||||||
if v is not None:
|
if v is not None:
|
||||||
return self.tx_zip(k, v, vn, rem, [], self.args.ed)
|
return self.tx_zip(k, v, vn, rem, [], self.args.ed)
|
||||||
|
|
||||||
fsroot, vfs_ls, vfs_virt = vn.ls(rem, self.uname, not self.args.no_scandir)
|
fsroot, vfs_ls, vfs_virt = vn.ls(
|
||||||
|
rem, self.uname, not self.args.no_scandir, incl_wo=True
|
||||||
|
)
|
||||||
stats = {k: v for k, v in vfs_ls}
|
stats = {k: v for k, v in vfs_ls}
|
||||||
vfs_ls = [x[0] for x in vfs_ls]
|
vfs_ls = [x[0] for x in vfs_ls]
|
||||||
vfs_ls.extend(vfs_virt.keys())
|
vfs_ls.extend(vfs_virt.keys())
|
||||||
@@ -1484,7 +1697,7 @@ class HttpCli(object):
|
|||||||
icur = None
|
icur = None
|
||||||
if "e2t" in vn.flags:
|
if "e2t" in vn.flags:
|
||||||
idx = self.conn.get_u2idx()
|
idx = self.conn.get_u2idx()
|
||||||
icur = idx.get_cur(vn.realpath)
|
icur = idx.get_cur(dbv.realpath)
|
||||||
|
|
||||||
dirs = []
|
dirs = []
|
||||||
files = []
|
files = []
|
||||||
@@ -1552,6 +1765,9 @@ class HttpCli(object):
|
|||||||
rd = f["rd"]
|
rd = f["rd"]
|
||||||
del f["rd"]
|
del f["rd"]
|
||||||
if icur:
|
if icur:
|
||||||
|
if vn != dbv:
|
||||||
|
_, rd = vn.get_dbv(rd)
|
||||||
|
|
||||||
q = "select w from up where rd = ? and fn = ?"
|
q = "select w from up where rd = ? and fn = ?"
|
||||||
try:
|
try:
|
||||||
r = icur.execute(q, (rd, fn)).fetchone()
|
r = icur.execute(q, (rd, fn)).fetchone()
|
||||||
@@ -1582,15 +1798,23 @@ class HttpCli(object):
|
|||||||
ls_ret["files"] = files
|
ls_ret["files"] = files
|
||||||
ls_ret["taglist"] = taglist
|
ls_ret["taglist"] = taglist
|
||||||
ret = json.dumps(ls_ret)
|
ret = json.dumps(ls_ret)
|
||||||
self.reply(ret.encode("utf-8", "replace"), mime="application/json")
|
self.reply(
|
||||||
|
ret.encode("utf-8", "replace"),
|
||||||
|
mime="application/json",
|
||||||
|
headers=NO_STORE,
|
||||||
|
)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
j2a["files"] = dirs + files
|
j2a["files"] = dirs + files
|
||||||
j2a["logues"] = logues
|
j2a["logues"] = logues
|
||||||
j2a["taglist"] = taglist
|
j2a["taglist"] = taglist
|
||||||
|
|
||||||
if "mte" in vn.flags:
|
if "mte" in vn.flags:
|
||||||
j2a["tag_order"] = json.dumps(vn.flags["mte"].split(","))
|
j2a["tag_order"] = json.dumps(vn.flags["mte"].split(","))
|
||||||
|
|
||||||
|
if self.args.css_browser:
|
||||||
|
j2a["css"] = self.args.css_browser
|
||||||
|
|
||||||
html = self.j2(tpl, **j2a)
|
html = self.j2(tpl, **j2a)
|
||||||
self.reply(html.encode("utf-8", "replace"))
|
self.reply(html.encode("utf-8", "replace"), headers=NO_STORE)
|
||||||
return True
|
return True
|
||||||
|
|||||||
@@ -17,6 +17,9 @@ from .__init__ import E
|
|||||||
from .util import Unrecv
|
from .util import Unrecv
|
||||||
from .httpcli import HttpCli
|
from .httpcli import HttpCli
|
||||||
from .u2idx import U2idx
|
from .u2idx import U2idx
|
||||||
|
from .th_cli import ThumbCli
|
||||||
|
from .th_srv import HAVE_PIL
|
||||||
|
from .ico import Ico
|
||||||
|
|
||||||
|
|
||||||
class HttpConn(object):
|
class HttpConn(object):
|
||||||
@@ -31,9 +34,14 @@ class HttpConn(object):
|
|||||||
self.hsrv = hsrv
|
self.hsrv = hsrv
|
||||||
|
|
||||||
self.args = hsrv.args
|
self.args = hsrv.args
|
||||||
self.auth = hsrv.auth
|
self.asrv = hsrv.asrv
|
||||||
|
self.is_mp = hsrv.is_mp
|
||||||
self.cert_path = hsrv.cert_path
|
self.cert_path = hsrv.cert_path
|
||||||
|
|
||||||
|
enth = HAVE_PIL and not self.args.no_thumb
|
||||||
|
self.thumbcli = ThumbCli(hsrv.broker) if enth else None
|
||||||
|
self.ico = Ico(self.args)
|
||||||
|
|
||||||
self.t0 = time.time()
|
self.t0 = time.time()
|
||||||
self.nbyte = 0
|
self.nbyte = 0
|
||||||
self.workload = 0
|
self.workload = 0
|
||||||
@@ -63,7 +71,7 @@ class HttpConn(object):
|
|||||||
|
|
||||||
def get_u2idx(self):
|
def get_u2idx(self):
|
||||||
if not self.u2idx:
|
if not self.u2idx:
|
||||||
self.u2idx = U2idx(self.args, self.log_func)
|
self.u2idx = U2idx(self)
|
||||||
|
|
||||||
return self.u2idx
|
return self.u2idx
|
||||||
|
|
||||||
@@ -167,6 +175,11 @@ class HttpConn(object):
|
|||||||
self.sr = Unrecv(self.s)
|
self.sr = Unrecv(self.s)
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
|
if self.is_mp:
|
||||||
|
self.workload += 50
|
||||||
|
if self.workload >= 2 ** 31:
|
||||||
|
self.workload = 100
|
||||||
|
|
||||||
cli = HttpCli(self)
|
cli = HttpCli(self)
|
||||||
if not cli.run():
|
if not cli.run():
|
||||||
return
|
return
|
||||||
|
|||||||
@@ -25,8 +25,8 @@ except ImportError:
|
|||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
from .__init__ import E, MACOS
|
from .__init__ import E, MACOS
|
||||||
from .httpconn import HttpConn
|
|
||||||
from .authsrv import AuthSrv
|
from .authsrv import AuthSrv
|
||||||
|
from .httpconn import HttpConn
|
||||||
|
|
||||||
|
|
||||||
class HttpSrv(object):
|
class HttpSrv(object):
|
||||||
@@ -35,10 +35,12 @@ class HttpSrv(object):
|
|||||||
relying on MpSrv for performance (HttpSrv is just plain threads)
|
relying on MpSrv for performance (HttpSrv is just plain threads)
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, broker):
|
def __init__(self, broker, is_mp=False):
|
||||||
self.broker = broker
|
self.broker = broker
|
||||||
|
self.is_mp = is_mp
|
||||||
self.args = broker.args
|
self.args = broker.args
|
||||||
self.log = broker.log
|
self.log = broker.log
|
||||||
|
self.asrv = broker.asrv
|
||||||
|
|
||||||
self.disconnect_func = None
|
self.disconnect_func = None
|
||||||
self.mutex = threading.Lock()
|
self.mutex = threading.Lock()
|
||||||
@@ -46,7 +48,6 @@ class HttpSrv(object):
|
|||||||
self.clients = {}
|
self.clients = {}
|
||||||
self.workload = 0
|
self.workload = 0
|
||||||
self.workload_thr_alive = False
|
self.workload_thr_alive = False
|
||||||
self.auth = AuthSrv(self.args, self.log)
|
|
||||||
|
|
||||||
env = jinja2.Environment()
|
env = jinja2.Environment()
|
||||||
env.loader = jinja2.FileSystemLoader(os.path.join(E.mod, "web"))
|
env.loader = jinja2.FileSystemLoader(os.path.join(E.mod, "web"))
|
||||||
@@ -66,7 +67,11 @@ class HttpSrv(object):
|
|||||||
if self.args.log_conn:
|
if self.args.log_conn:
|
||||||
self.log("%s %s" % addr, "|%sC-cthr" % ("-" * 5,), c="1;30")
|
self.log("%s %s" % addr, "|%sC-cthr" % ("-" * 5,), c="1;30")
|
||||||
|
|
||||||
thr = threading.Thread(target=self.thr_client, args=(sck, addr))
|
thr = threading.Thread(
|
||||||
|
target=self.thr_client,
|
||||||
|
args=(sck, addr),
|
||||||
|
name="httpsrv-{}-{}".format(addr[0].split(".", 2)[-1][-6:], addr[1]),
|
||||||
|
)
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
@@ -84,11 +89,14 @@ class HttpSrv(object):
|
|||||||
cli = HttpConn(sck, addr, self)
|
cli = HttpConn(sck, addr, self)
|
||||||
with self.mutex:
|
with self.mutex:
|
||||||
self.clients[cli] = 0
|
self.clients[cli] = 0
|
||||||
self.workload += 50
|
|
||||||
|
|
||||||
|
if self.is_mp:
|
||||||
|
self.workload += 50
|
||||||
if not self.workload_thr_alive:
|
if not self.workload_thr_alive:
|
||||||
self.workload_thr_alive = True
|
self.workload_thr_alive = True
|
||||||
thr = threading.Thread(target=self.thr_workload)
|
thr = threading.Thread(
|
||||||
|
target=self.thr_workload, name="httpsrv-workload"
|
||||||
|
)
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
@@ -99,6 +107,7 @@ class HttpSrv(object):
|
|||||||
cli.run()
|
cli.run()
|
||||||
|
|
||||||
finally:
|
finally:
|
||||||
|
sck = cli.s
|
||||||
if self.args.log_conn:
|
if self.args.log_conn:
|
||||||
self.log("%s %s" % addr, "|%sC-cdone" % ("-" * 7,), c="1;30")
|
self.log("%s %s" % addr, "|%sC-cdone" % ("-" * 7,), c="1;30")
|
||||||
|
|
||||||
|
|||||||
42
copyparty/ico.py
Normal file
42
copyparty/ico.py
Normal file
@@ -0,0 +1,42 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
|
import hashlib
|
||||||
|
import colorsys
|
||||||
|
|
||||||
|
from .__init__ import PY2
|
||||||
|
|
||||||
|
|
||||||
|
class Ico(object):
|
||||||
|
def __init__(self, args):
|
||||||
|
self.args = args
|
||||||
|
|
||||||
|
def get(self, ext, as_thumb):
|
||||||
|
"""placeholder to make thumbnails not break"""
|
||||||
|
|
||||||
|
h = hashlib.md5(ext.encode("utf-8")).digest()[:2]
|
||||||
|
if PY2:
|
||||||
|
h = [ord(x) for x in h]
|
||||||
|
|
||||||
|
c1 = colorsys.hsv_to_rgb(h[0] / 256.0, 1, 0.3)
|
||||||
|
c2 = colorsys.hsv_to_rgb(h[0] / 256.0, 1, 1)
|
||||||
|
c = list(c1) + list(c2)
|
||||||
|
c = [int(x * 255) for x in c]
|
||||||
|
c = "".join(["{:02x}".format(x) for x in c])
|
||||||
|
|
||||||
|
h = 30
|
||||||
|
if not self.args.th_no_crop and as_thumb:
|
||||||
|
w, h = self.args.th_size.split("x")
|
||||||
|
h = int(100 / (float(w) / float(h)))
|
||||||
|
|
||||||
|
svg = """\
|
||||||
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
<svg version="1.1" viewBox="0 0 100 {}" xmlns="http://www.w3.org/2000/svg"><g>
|
||||||
|
<rect width="100%" height="100%" fill="#{}" />
|
||||||
|
<text x="50%" y="50%" dominant-baseline="middle" text-anchor="middle" xml:space="preserve"
|
||||||
|
fill="#{}" font-family="monospace" font-size="14px" style="letter-spacing:.5px">{}</text>
|
||||||
|
</g></svg>
|
||||||
|
"""
|
||||||
|
svg = svg.format(h, c[:6], c[6:], ext).encode("utf-8")
|
||||||
|
|
||||||
|
return ["image/svg+xml", svg]
|
||||||
@@ -1,9 +1,9 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
from __future__ import print_function, unicode_literals
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
import re
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
import json
|
||||||
import shutil
|
import shutil
|
||||||
import subprocess as sp
|
import subprocess as sp
|
||||||
|
|
||||||
@@ -14,6 +14,204 @@ if not PY2:
|
|||||||
unicode = str
|
unicode = str
|
||||||
|
|
||||||
|
|
||||||
|
def have_ff(cmd):
|
||||||
|
if PY2:
|
||||||
|
cmd = (cmd + " -version").encode("ascii").split(b" ")
|
||||||
|
try:
|
||||||
|
sp.Popen(cmd, stdout=sp.PIPE, stderr=sp.PIPE).communicate()
|
||||||
|
return True
|
||||||
|
except:
|
||||||
|
return False
|
||||||
|
else:
|
||||||
|
return bool(shutil.which(cmd))
|
||||||
|
|
||||||
|
|
||||||
|
HAVE_FFMPEG = have_ff("ffmpeg")
|
||||||
|
HAVE_FFPROBE = have_ff("ffprobe")
|
||||||
|
|
||||||
|
|
||||||
|
class MParser(object):
|
||||||
|
def __init__(self, cmdline):
|
||||||
|
self.tag, args = cmdline.split("=", 1)
|
||||||
|
self.tags = self.tag.split(",")
|
||||||
|
|
||||||
|
self.timeout = 30
|
||||||
|
self.force = False
|
||||||
|
self.audio = "y"
|
||||||
|
self.ext = []
|
||||||
|
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
bp = os.path.expanduser(args)
|
||||||
|
if os.path.exists(bp):
|
||||||
|
self.bin = bp
|
||||||
|
return
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
arg, args = args.split(",", 1)
|
||||||
|
arg = arg.lower()
|
||||||
|
|
||||||
|
if arg.startswith("a"):
|
||||||
|
self.audio = arg[1:] # [r]equire [n]ot [d]ontcare
|
||||||
|
continue
|
||||||
|
|
||||||
|
if arg == "f":
|
||||||
|
self.force = True
|
||||||
|
continue
|
||||||
|
|
||||||
|
if arg.startswith("t"):
|
||||||
|
self.timeout = int(arg[1:])
|
||||||
|
continue
|
||||||
|
|
||||||
|
if arg.startswith("e"):
|
||||||
|
self.ext.append(arg[1:])
|
||||||
|
continue
|
||||||
|
|
||||||
|
raise Exception()
|
||||||
|
|
||||||
|
|
||||||
|
def ffprobe(abspath):
|
||||||
|
cmd = [
|
||||||
|
b"ffprobe",
|
||||||
|
b"-hide_banner",
|
||||||
|
b"-show_streams",
|
||||||
|
b"-show_format",
|
||||||
|
b"--",
|
||||||
|
fsenc(abspath),
|
||||||
|
]
|
||||||
|
p = sp.Popen(cmd, stdout=sp.PIPE, stderr=sp.PIPE)
|
||||||
|
r = p.communicate()
|
||||||
|
txt = r[0].decode("utf-8", "replace")
|
||||||
|
return parse_ffprobe(txt)
|
||||||
|
|
||||||
|
|
||||||
|
def parse_ffprobe(txt):
|
||||||
|
"""ffprobe -show_format -show_streams"""
|
||||||
|
streams = []
|
||||||
|
fmt = {}
|
||||||
|
g = None
|
||||||
|
for ln in [x.rstrip("\r") for x in txt.split("\n")]:
|
||||||
|
try:
|
||||||
|
k, v = ln.split("=", 1)
|
||||||
|
g[k] = v
|
||||||
|
continue
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
if ln == "[STREAM]":
|
||||||
|
g = {}
|
||||||
|
streams.append(g)
|
||||||
|
|
||||||
|
if ln == "[FORMAT]":
|
||||||
|
g = {"codec_type": "format"} # heh
|
||||||
|
fmt = g
|
||||||
|
|
||||||
|
streams = [fmt] + streams
|
||||||
|
ret = {} # processed
|
||||||
|
md = {} # raw tags
|
||||||
|
|
||||||
|
have = {}
|
||||||
|
for strm in streams:
|
||||||
|
typ = strm.get("codec_type")
|
||||||
|
if typ in have:
|
||||||
|
continue
|
||||||
|
|
||||||
|
have[typ] = True
|
||||||
|
kvm = []
|
||||||
|
|
||||||
|
if typ == "audio":
|
||||||
|
kvm = [
|
||||||
|
["codec_name", "ac"],
|
||||||
|
["channel_layout", "chs"],
|
||||||
|
["sample_rate", ".hz"],
|
||||||
|
["bit_rate", ".aq"],
|
||||||
|
["duration", ".dur"],
|
||||||
|
]
|
||||||
|
|
||||||
|
if typ == "video":
|
||||||
|
if strm.get("DISPOSITION:attached_pic") == "1" or fmt.get(
|
||||||
|
"format_name"
|
||||||
|
) in ["mp3", "ogg", "flac"]:
|
||||||
|
continue
|
||||||
|
|
||||||
|
kvm = [
|
||||||
|
["codec_name", "vc"],
|
||||||
|
["pix_fmt", "pixfmt"],
|
||||||
|
["r_frame_rate", ".fps"],
|
||||||
|
["bit_rate", ".vq"],
|
||||||
|
["width", ".resw"],
|
||||||
|
["height", ".resh"],
|
||||||
|
["duration", ".dur"],
|
||||||
|
]
|
||||||
|
|
||||||
|
if typ == "format":
|
||||||
|
kvm = [["duration", ".dur"], ["bit_rate", ".q"]]
|
||||||
|
|
||||||
|
for sk, rk in kvm:
|
||||||
|
v = strm.get(sk)
|
||||||
|
if v is None:
|
||||||
|
continue
|
||||||
|
|
||||||
|
if rk.startswith("."):
|
||||||
|
try:
|
||||||
|
v = float(v)
|
||||||
|
v2 = ret.get(rk)
|
||||||
|
if v2 is None or v > v2:
|
||||||
|
ret[rk] = v
|
||||||
|
except:
|
||||||
|
# sqlite doesnt care but the code below does
|
||||||
|
if v not in ["N/A"]:
|
||||||
|
ret[rk] = v
|
||||||
|
else:
|
||||||
|
ret[rk] = v
|
||||||
|
|
||||||
|
if ret.get("vc") == "ansi": # shellscript
|
||||||
|
return {}, {}
|
||||||
|
|
||||||
|
for strm in streams:
|
||||||
|
for k, v in strm.items():
|
||||||
|
if not k.startswith("TAG:"):
|
||||||
|
continue
|
||||||
|
|
||||||
|
k = k[4:].strip()
|
||||||
|
v = v.strip()
|
||||||
|
if k and v:
|
||||||
|
md[k] = [v]
|
||||||
|
|
||||||
|
for k in [".q", ".vq", ".aq"]:
|
||||||
|
if k in ret:
|
||||||
|
ret[k] /= 1000 # bit_rate=320000
|
||||||
|
|
||||||
|
for k in [".q", ".vq", ".aq", ".resw", ".resh"]:
|
||||||
|
if k in ret:
|
||||||
|
ret[k] = int(ret[k])
|
||||||
|
|
||||||
|
if ".fps" in ret:
|
||||||
|
fps = ret[".fps"]
|
||||||
|
if "/" in fps:
|
||||||
|
fa, fb = fps.split("/")
|
||||||
|
fps = int(fa) * 1.0 / int(fb)
|
||||||
|
|
||||||
|
if fps < 1000 and fmt.get("format_name") not in ["image2", "png_pipe"]:
|
||||||
|
ret[".fps"] = round(fps, 3)
|
||||||
|
else:
|
||||||
|
del ret[".fps"]
|
||||||
|
|
||||||
|
if ".dur" in ret:
|
||||||
|
if ret[".dur"] < 0.1:
|
||||||
|
del ret[".dur"]
|
||||||
|
if ".q" in ret:
|
||||||
|
del ret[".q"]
|
||||||
|
|
||||||
|
if ".resw" in ret and ".resh" in ret:
|
||||||
|
ret["res"] = "{}x{}".format(ret[".resw"], ret[".resh"])
|
||||||
|
|
||||||
|
ret = {k: [0, v] for k, v in ret.items()}
|
||||||
|
|
||||||
|
return ret, md
|
||||||
|
|
||||||
|
|
||||||
class MTag(object):
|
class MTag(object):
|
||||||
def __init__(self, log_func, args):
|
def __init__(self, log_func, args):
|
||||||
self.log_func = log_func
|
self.log_func = log_func
|
||||||
@@ -35,15 +233,7 @@ class MTag(object):
|
|||||||
self.get = self.get_ffprobe
|
self.get = self.get_ffprobe
|
||||||
self.prefer_mt = True
|
self.prefer_mt = True
|
||||||
# about 20x slower
|
# about 20x slower
|
||||||
if PY2:
|
self.usable = HAVE_FFPROBE
|
||||||
cmd = [b"ffprobe", b"-version"]
|
|
||||||
try:
|
|
||||||
sp.Popen(cmd, stdout=sp.PIPE, stderr=sp.PIPE)
|
|
||||||
except:
|
|
||||||
self.usable = False
|
|
||||||
else:
|
|
||||||
if not shutil.which("ffprobe"):
|
|
||||||
self.usable = False
|
|
||||||
|
|
||||||
if self.usable and WINDOWS and sys.version_info < (3, 8):
|
if self.usable and WINDOWS and sys.version_info < (3, 8):
|
||||||
self.usable = False
|
self.usable = False
|
||||||
@@ -52,8 +242,10 @@ class MTag(object):
|
|||||||
self.log(msg, c=1)
|
self.log(msg, c=1)
|
||||||
|
|
||||||
if not self.usable:
|
if not self.usable:
|
||||||
msg = "need mutagen{} to read media tags so please run this:\n {} -m pip install --user mutagen"
|
msg = "need mutagen{} to read media tags so please run this:\n{}{} -m pip install --user mutagen\n"
|
||||||
self.log(msg.format(or_ffprobe, os.path.basename(sys.executable)), c=1)
|
self.log(
|
||||||
|
msg.format(or_ffprobe, " " * 37, os.path.basename(sys.executable)), c=1
|
||||||
|
)
|
||||||
return
|
return
|
||||||
|
|
||||||
# https://picard-docs.musicbrainz.org/downloads/MusicBrainz_Picard_Tag_Map.html
|
# https://picard-docs.musicbrainz.org/downloads/MusicBrainz_Picard_Tag_Map.html
|
||||||
@@ -201,7 +393,7 @@ class MTag(object):
|
|||||||
import mutagen
|
import mutagen
|
||||||
|
|
||||||
try:
|
try:
|
||||||
md = mutagen.File(abspath, easy=True)
|
md = mutagen.File(fsenc(abspath), easy=True)
|
||||||
x = md.info.length
|
x = md.info.length
|
||||||
except Exception as ex:
|
except Exception as ex:
|
||||||
return {}
|
return {}
|
||||||
@@ -212,7 +404,7 @@ class MTag(object):
|
|||||||
try:
|
try:
|
||||||
q = int(md.info.bitrate / 1024)
|
q = int(md.info.bitrate / 1024)
|
||||||
except:
|
except:
|
||||||
q = int((os.path.getsize(abspath) / dur) / 128)
|
q = int((os.path.getsize(fsenc(abspath)) / dur) / 128)
|
||||||
|
|
||||||
ret[".dur"] = [0, dur]
|
ret[".dur"] = [0, dur]
|
||||||
ret[".q"] = [0, q]
|
ret[".q"] = [0, q]
|
||||||
@@ -222,101 +414,7 @@ class MTag(object):
|
|||||||
return self.normalize_tags(ret, md)
|
return self.normalize_tags(ret, md)
|
||||||
|
|
||||||
def get_ffprobe(self, abspath):
|
def get_ffprobe(self, abspath):
|
||||||
cmd = [b"ffprobe", b"-hide_banner", b"--", fsenc(abspath)]
|
ret, md = ffprobe(abspath)
|
||||||
p = sp.Popen(cmd, stdout=sp.PIPE, stderr=sp.PIPE)
|
|
||||||
r = p.communicate()
|
|
||||||
txt = r[1].decode("utf-8", "replace")
|
|
||||||
txt = [x.rstrip("\r") for x in txt.split("\n")]
|
|
||||||
|
|
||||||
"""
|
|
||||||
note:
|
|
||||||
tags which contain newline will be truncated on first \n,
|
|
||||||
ffprobe emits \n and spacepads the : to align visually
|
|
||||||
note:
|
|
||||||
the Stream ln always mentions Audio: if audio
|
|
||||||
the Stream ln usually has kb/s, is more accurate
|
|
||||||
the Duration ln always has kb/s
|
|
||||||
the Metadata: after Chapter may contain BPM info,
|
|
||||||
title : Tempo: 126.0
|
|
||||||
|
|
||||||
Input #0, wav,
|
|
||||||
Metadata:
|
|
||||||
date : <OK>
|
|
||||||
Duration:
|
|
||||||
Chapter #
|
|
||||||
Metadata:
|
|
||||||
title : <NG>
|
|
||||||
|
|
||||||
Input #0, mp3,
|
|
||||||
Metadata:
|
|
||||||
album : <OK>
|
|
||||||
Duration:
|
|
||||||
Stream #0:0: Audio:
|
|
||||||
Stream #0:1: Video:
|
|
||||||
Metadata:
|
|
||||||
comment : <NG>
|
|
||||||
"""
|
|
||||||
|
|
||||||
ptn_md_beg = re.compile("^( +)Metadata:$")
|
|
||||||
ptn_md_kv = re.compile("^( +)([^:]+) *: (.*)")
|
|
||||||
ptn_dur = re.compile("^ *Duration: ([^ ]+)(, |$)")
|
|
||||||
ptn_br1 = re.compile("^ *Duration: .*, bitrate: ([0-9]+) kb/s(, |$)")
|
|
||||||
ptn_br2 = re.compile("^ *Stream.*: Audio:.* ([0-9]+) kb/s(, |$)")
|
|
||||||
ptn_audio = re.compile("^ *Stream .*: Audio: ")
|
|
||||||
ptn_au_parent = re.compile("^ *(Input #|Stream .*: Audio: )")
|
|
||||||
|
|
||||||
ret = {}
|
|
||||||
md = {}
|
|
||||||
in_md = False
|
|
||||||
is_audio = False
|
|
||||||
au_parent = False
|
|
||||||
for ln in txt:
|
|
||||||
m = ptn_md_kv.match(ln)
|
|
||||||
if m and in_md and len(m.group(1)) == in_md:
|
|
||||||
_, k, v = [x.strip() for x in m.groups()]
|
|
||||||
if k != "" and v != "":
|
|
||||||
md[k] = [v]
|
|
||||||
continue
|
|
||||||
else:
|
|
||||||
in_md = False
|
|
||||||
|
|
||||||
m = ptn_md_beg.match(ln)
|
|
||||||
if m and au_parent:
|
|
||||||
in_md = len(m.group(1)) + 2
|
|
||||||
continue
|
|
||||||
|
|
||||||
au_parent = bool(ptn_au_parent.search(ln))
|
|
||||||
|
|
||||||
if ptn_audio.search(ln):
|
|
||||||
is_audio = True
|
|
||||||
|
|
||||||
m = ptn_dur.search(ln)
|
|
||||||
if m:
|
|
||||||
sec = 0
|
|
||||||
tstr = m.group(1)
|
|
||||||
if tstr.lower() != "n/a":
|
|
||||||
try:
|
|
||||||
tf = tstr.split(",")[0].split(".")[0].split(":")
|
|
||||||
for f in tf:
|
|
||||||
sec *= 60
|
|
||||||
sec += int(f)
|
|
||||||
except:
|
|
||||||
self.log("invalid timestr from ffprobe: [{}]".format(tstr), c=3)
|
|
||||||
|
|
||||||
ret[".dur"] = sec
|
|
||||||
m = ptn_br1.search(ln)
|
|
||||||
if m:
|
|
||||||
ret[".q"] = m.group(1)
|
|
||||||
|
|
||||||
m = ptn_br2.search(ln)
|
|
||||||
if m:
|
|
||||||
ret[".q"] = m.group(1)
|
|
||||||
|
|
||||||
if not is_audio:
|
|
||||||
return {}
|
|
||||||
|
|
||||||
ret = {k: [0, v] for k, v in ret.items()}
|
|
||||||
|
|
||||||
return self.normalize_tags(ret, md)
|
return self.normalize_tags(ret, md)
|
||||||
|
|
||||||
def get_bin(self, parsers, abspath):
|
def get_bin(self, parsers, abspath):
|
||||||
@@ -327,10 +425,10 @@ class MTag(object):
|
|||||||
env["PYTHONPATH"] = pypath
|
env["PYTHONPATH"] = pypath
|
||||||
|
|
||||||
ret = {}
|
ret = {}
|
||||||
for tagname, (binpath, timeout) in parsers.items():
|
for tagname, mp in parsers.items():
|
||||||
try:
|
try:
|
||||||
cmd = [sys.executable, binpath, abspath]
|
cmd = [sys.executable, mp.bin, abspath]
|
||||||
args = {"env": env, "timeout": timeout}
|
args = {"env": env, "timeout": mp.timeout}
|
||||||
|
|
||||||
if WINDOWS:
|
if WINDOWS:
|
||||||
args["creationflags"] = 0x4000
|
args["creationflags"] = 0x4000
|
||||||
@@ -339,8 +437,16 @@ class MTag(object):
|
|||||||
|
|
||||||
cmd = [fsenc(x) for x in cmd]
|
cmd = [fsenc(x) for x in cmd]
|
||||||
v = sp.check_output(cmd, **args).strip()
|
v = sp.check_output(cmd, **args).strip()
|
||||||
if v:
|
if not v:
|
||||||
|
continue
|
||||||
|
|
||||||
|
if "," not in tagname:
|
||||||
ret[tagname] = v.decode("utf-8")
|
ret[tagname] = v.decode("utf-8")
|
||||||
|
else:
|
||||||
|
v = json.loads(v)
|
||||||
|
for tag in tagname.split(","):
|
||||||
|
if tag and tag in v:
|
||||||
|
ret[tag] = v[tag]
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|||||||
@@ -1,3 +1,6 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import tarfile
|
import tarfile
|
||||||
import threading
|
import threading
|
||||||
@@ -42,7 +45,7 @@ class StreamTar(object):
|
|||||||
fmt = tarfile.GNU_FORMAT
|
fmt = tarfile.GNU_FORMAT
|
||||||
self.tar = tarfile.open(fileobj=self.qfile, mode="w|", format=fmt)
|
self.tar = tarfile.open(fileobj=self.qfile, mode="w|", format=fmt)
|
||||||
|
|
||||||
w = threading.Thread(target=self._gen)
|
w = threading.Thread(target=self._gen, name="star-gen")
|
||||||
w.daemon = True
|
w.daemon = True
|
||||||
w.start()
|
w.start()
|
||||||
|
|
||||||
|
|||||||
@@ -1,3 +1,6 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import time
|
import time
|
||||||
import tempfile
|
import tempfile
|
||||||
|
|||||||
@@ -2,6 +2,7 @@
|
|||||||
from __future__ import print_function, unicode_literals
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
import re
|
import re
|
||||||
|
import os
|
||||||
import sys
|
import sys
|
||||||
import time
|
import time
|
||||||
import threading
|
import threading
|
||||||
@@ -9,9 +10,11 @@ from datetime import datetime, timedelta
|
|||||||
import calendar
|
import calendar
|
||||||
|
|
||||||
from .__init__ import PY2, WINDOWS, MACOS, VT100
|
from .__init__ import PY2, WINDOWS, MACOS, VT100
|
||||||
|
from .util import mp
|
||||||
|
from .authsrv import AuthSrv
|
||||||
from .tcpsrv import TcpSrv
|
from .tcpsrv import TcpSrv
|
||||||
from .up2k import Up2k
|
from .up2k import Up2k
|
||||||
from .util import mp
|
from .th_srv import ThumbSrv, HAVE_PIL, HAVE_WEBP
|
||||||
|
|
||||||
|
|
||||||
class SvcHub(object):
|
class SvcHub(object):
|
||||||
@@ -35,9 +38,28 @@ class SvcHub(object):
|
|||||||
self.log = self._log_disabled if args.q else self._log_enabled
|
self.log = self._log_disabled if args.q else self._log_enabled
|
||||||
|
|
||||||
# initiate all services to manage
|
# initiate all services to manage
|
||||||
|
self.asrv = AuthSrv(self.args, self.log, False)
|
||||||
|
if args.ls:
|
||||||
|
self.asrv.dbg_ls()
|
||||||
|
|
||||||
self.tcpsrv = TcpSrv(self)
|
self.tcpsrv = TcpSrv(self)
|
||||||
self.up2k = Up2k(self)
|
self.up2k = Up2k(self)
|
||||||
|
|
||||||
|
self.thumbsrv = None
|
||||||
|
if not args.no_thumb:
|
||||||
|
if HAVE_PIL:
|
||||||
|
if not HAVE_WEBP:
|
||||||
|
args.th_no_webp = True
|
||||||
|
msg = "setting --th-no-webp because either libwebp is not available or your Pillow is too old"
|
||||||
|
self.log("thumb", msg, c=3)
|
||||||
|
|
||||||
|
self.thumbsrv = ThumbSrv(self)
|
||||||
|
else:
|
||||||
|
msg = "need Pillow to create thumbnails; for example:\n{}{} -m pip install --user Pillow\n"
|
||||||
|
self.log(
|
||||||
|
"thumb", msg.format(" " * 37, os.path.basename(sys.executable)), c=3
|
||||||
|
)
|
||||||
|
|
||||||
# decide which worker impl to use
|
# decide which worker impl to use
|
||||||
if self.check_mp_enable():
|
if self.check_mp_enable():
|
||||||
from .broker_mp import BrokerMp as Broker
|
from .broker_mp import BrokerMp as Broker
|
||||||
@@ -48,7 +70,7 @@ class SvcHub(object):
|
|||||||
self.broker = Broker(self)
|
self.broker = Broker(self)
|
||||||
|
|
||||||
def run(self):
|
def run(self):
|
||||||
thr = threading.Thread(target=self.tcpsrv.run)
|
thr = threading.Thread(target=self.tcpsrv.run, name="svchub-main")
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
@@ -63,7 +85,20 @@ class SvcHub(object):
|
|||||||
|
|
||||||
self.tcpsrv.shutdown()
|
self.tcpsrv.shutdown()
|
||||||
self.broker.shutdown()
|
self.broker.shutdown()
|
||||||
print("nailed it")
|
if self.thumbsrv:
|
||||||
|
self.thumbsrv.shutdown()
|
||||||
|
|
||||||
|
for n in range(200): # 10s
|
||||||
|
time.sleep(0.05)
|
||||||
|
if self.thumbsrv.stopped():
|
||||||
|
break
|
||||||
|
|
||||||
|
if n == 3:
|
||||||
|
print("waiting for thumbsrv (10sec)...")
|
||||||
|
|
||||||
|
print("nailed it", end="")
|
||||||
|
finally:
|
||||||
|
print("\033[0m")
|
||||||
|
|
||||||
def _log_disabled(self, src, msg, c=0):
|
def _log_disabled(self, src, msg, c=0):
|
||||||
pass
|
pass
|
||||||
|
|||||||
@@ -1,3 +1,6 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import time
|
import time
|
||||||
import zlib
|
import zlib
|
||||||
|
|||||||
55
copyparty/th_cli.py
Normal file
55
copyparty/th_cli.py
Normal file
@@ -0,0 +1,55 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
|
import os
|
||||||
|
|
||||||
|
from .util import Cooldown
|
||||||
|
from .th_srv import thumb_path, THUMBABLE, FMT_FF
|
||||||
|
|
||||||
|
|
||||||
|
class ThumbCli(object):
|
||||||
|
def __init__(self, broker):
|
||||||
|
self.broker = broker
|
||||||
|
self.args = broker.args
|
||||||
|
self.asrv = broker.asrv
|
||||||
|
|
||||||
|
# cache on both sides for less broker spam
|
||||||
|
self.cooldown = Cooldown(self.args.th_poke)
|
||||||
|
|
||||||
|
def get(self, ptop, rem, mtime, fmt):
|
||||||
|
ext = rem.rsplit(".")[-1].lower()
|
||||||
|
if ext not in THUMBABLE:
|
||||||
|
return None
|
||||||
|
|
||||||
|
is_vid = ext in FMT_FF
|
||||||
|
if is_vid and self.args.no_vthumb:
|
||||||
|
return None
|
||||||
|
|
||||||
|
if fmt == "j" and self.args.th_no_jpg:
|
||||||
|
fmt = "w"
|
||||||
|
|
||||||
|
if fmt == "w":
|
||||||
|
if self.args.th_no_webp or (is_vid and self.args.th_ff_jpg):
|
||||||
|
fmt = "j"
|
||||||
|
|
||||||
|
histpath = self.asrv.vfs.histtab[ptop]
|
||||||
|
tpath = thumb_path(histpath, rem, mtime, fmt)
|
||||||
|
ret = None
|
||||||
|
try:
|
||||||
|
st = os.stat(tpath)
|
||||||
|
if st.st_size:
|
||||||
|
ret = tpath
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
if ret:
|
||||||
|
tdir = os.path.dirname(tpath)
|
||||||
|
if self.cooldown.poke(tdir):
|
||||||
|
self.broker.put(False, "thumbsrv.poke", tdir)
|
||||||
|
|
||||||
|
return ret
|
||||||
|
|
||||||
|
x = self.broker.put(True, "thumbsrv.get", ptop, rem, mtime, fmt)
|
||||||
|
return x.get()
|
||||||
399
copyparty/th_srv.py
Normal file
399
copyparty/th_srv.py
Normal file
@@ -0,0 +1,399 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import print_function, unicode_literals
|
||||||
|
|
||||||
|
import os
|
||||||
|
import time
|
||||||
|
import shutil
|
||||||
|
import base64
|
||||||
|
import hashlib
|
||||||
|
import threading
|
||||||
|
import subprocess as sp
|
||||||
|
|
||||||
|
from .__init__ import PY2
|
||||||
|
from .util import fsenc, runcmd, Queue, Cooldown, BytesIO, min_ex
|
||||||
|
from .mtag import HAVE_FFMPEG, HAVE_FFPROBE, ffprobe
|
||||||
|
|
||||||
|
|
||||||
|
if not PY2:
|
||||||
|
unicode = str
|
||||||
|
|
||||||
|
|
||||||
|
HAVE_PIL = False
|
||||||
|
HAVE_HEIF = False
|
||||||
|
HAVE_AVIF = False
|
||||||
|
HAVE_WEBP = False
|
||||||
|
|
||||||
|
try:
|
||||||
|
from PIL import Image, ImageOps
|
||||||
|
|
||||||
|
HAVE_PIL = True
|
||||||
|
try:
|
||||||
|
Image.new("RGB", (2, 2)).save(BytesIO(), format="webp")
|
||||||
|
HAVE_WEBP = True
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
try:
|
||||||
|
from pyheif_pillow_opener import register_heif_opener
|
||||||
|
|
||||||
|
register_heif_opener()
|
||||||
|
HAVE_HEIF = True
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
try:
|
||||||
|
import pillow_avif
|
||||||
|
|
||||||
|
HAVE_AVIF = True
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
# https://pillow.readthedocs.io/en/stable/handbook/image-file-formats.html
|
||||||
|
# ffmpeg -formats
|
||||||
|
FMT_PIL = "bmp dib gif icns ico jpg jpeg jp2 jpx pcx png pbm pgm ppm pnm sgi tga tif tiff webp xbm dds xpm"
|
||||||
|
FMT_FF = "av1 asf avi flv m4v mkv mjpeg mjpg mpg mpeg mpg2 mpeg2 h264 avc h265 hevc mov 3gp mp4 ts mpegts nut ogv ogm rm vob webm wmv"
|
||||||
|
|
||||||
|
if HAVE_HEIF:
|
||||||
|
FMT_PIL += " heif heifs heic heics"
|
||||||
|
|
||||||
|
if HAVE_AVIF:
|
||||||
|
FMT_PIL += " avif avifs"
|
||||||
|
|
||||||
|
FMT_PIL, FMT_FF = [{x: True for x in y.split(" ") if x} for y in [FMT_PIL, FMT_FF]]
|
||||||
|
|
||||||
|
|
||||||
|
THUMBABLE = {}
|
||||||
|
|
||||||
|
if HAVE_PIL:
|
||||||
|
THUMBABLE.update(FMT_PIL)
|
||||||
|
|
||||||
|
if HAVE_FFMPEG and HAVE_FFPROBE:
|
||||||
|
THUMBABLE.update(FMT_FF)
|
||||||
|
|
||||||
|
|
||||||
|
def thumb_path(histpath, rem, mtime, fmt):
|
||||||
|
# base16 = 16 = 256
|
||||||
|
# b64-lc = 38 = 1444
|
||||||
|
# base64 = 64 = 4096
|
||||||
|
try:
|
||||||
|
rd, fn = rem.rsplit("/", 1)
|
||||||
|
except:
|
||||||
|
rd = ""
|
||||||
|
fn = rem
|
||||||
|
|
||||||
|
if rd:
|
||||||
|
h = hashlib.sha512(fsenc(rd)).digest()[:24]
|
||||||
|
b64 = base64.urlsafe_b64encode(h).decode("ascii")[:24]
|
||||||
|
rd = "{}/{}/".format(b64[:2], b64[2:4]).lower() + b64
|
||||||
|
else:
|
||||||
|
rd = "top"
|
||||||
|
|
||||||
|
# could keep original filenames but this is safer re pathlen
|
||||||
|
h = hashlib.sha512(fsenc(fn)).digest()[:24]
|
||||||
|
fn = base64.urlsafe_b64encode(h).decode("ascii")[:24]
|
||||||
|
|
||||||
|
return "{}/th/{}/{}.{:x}.{}".format(
|
||||||
|
histpath, rd, fn, int(mtime), "webp" if fmt == "w" else "jpg"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class ThumbSrv(object):
|
||||||
|
def __init__(self, hub):
|
||||||
|
self.hub = hub
|
||||||
|
self.asrv = hub.asrv
|
||||||
|
self.args = hub.args
|
||||||
|
self.log_func = hub.log
|
||||||
|
|
||||||
|
res = hub.args.th_size.split("x")
|
||||||
|
self.res = tuple([int(x) for x in res])
|
||||||
|
self.poke_cd = Cooldown(self.args.th_poke)
|
||||||
|
|
||||||
|
self.mutex = threading.Lock()
|
||||||
|
self.busy = {}
|
||||||
|
self.stopping = False
|
||||||
|
self.nthr = os.cpu_count() if hasattr(os, "cpu_count") else 4
|
||||||
|
self.q = Queue(self.nthr * 4)
|
||||||
|
for n in range(self.nthr):
|
||||||
|
t = threading.Thread(
|
||||||
|
target=self.worker, name="thumb-{}-{}".format(n, self.nthr)
|
||||||
|
)
|
||||||
|
t.daemon = True
|
||||||
|
t.start()
|
||||||
|
|
||||||
|
if not self.args.no_vthumb and (not HAVE_FFMPEG or not HAVE_FFPROBE):
|
||||||
|
missing = []
|
||||||
|
if not HAVE_FFMPEG:
|
||||||
|
missing.append("ffmpeg")
|
||||||
|
|
||||||
|
if not HAVE_FFPROBE:
|
||||||
|
missing.append("ffprobe")
|
||||||
|
|
||||||
|
msg = "cannot create video thumbnails because some of the required programs are not available: "
|
||||||
|
msg += ", ".join(missing)
|
||||||
|
self.log(msg, c=3)
|
||||||
|
|
||||||
|
t = threading.Thread(target=self.cleaner, name="thumb-cleaner")
|
||||||
|
t.daemon = True
|
||||||
|
t.start()
|
||||||
|
|
||||||
|
def log(self, msg, c=0):
|
||||||
|
self.log_func("thumb", msg, c)
|
||||||
|
|
||||||
|
def shutdown(self):
|
||||||
|
self.stopping = True
|
||||||
|
for _ in range(self.nthr):
|
||||||
|
self.q.put(None)
|
||||||
|
|
||||||
|
def stopped(self):
|
||||||
|
with self.mutex:
|
||||||
|
return not self.nthr
|
||||||
|
|
||||||
|
def get(self, ptop, rem, mtime, fmt):
|
||||||
|
histpath = self.asrv.vfs.histtab[ptop]
|
||||||
|
tpath = thumb_path(histpath, rem, mtime, fmt)
|
||||||
|
abspath = os.path.join(ptop, rem)
|
||||||
|
cond = threading.Condition()
|
||||||
|
with self.mutex:
|
||||||
|
try:
|
||||||
|
self.busy[tpath].append(cond)
|
||||||
|
self.log("wait {}".format(tpath))
|
||||||
|
except:
|
||||||
|
thdir = os.path.dirname(tpath)
|
||||||
|
try:
|
||||||
|
os.makedirs(thdir)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
inf_path = os.path.join(thdir, "dir.txt")
|
||||||
|
if not os.path.exists(inf_path):
|
||||||
|
with open(inf_path, "wb") as f:
|
||||||
|
f.write(fsenc(os.path.dirname(abspath)))
|
||||||
|
|
||||||
|
self.busy[tpath] = [cond]
|
||||||
|
self.q.put([abspath, tpath])
|
||||||
|
self.log("conv {} \033[0m{}".format(tpath, abspath), c=6)
|
||||||
|
|
||||||
|
while not self.stopping:
|
||||||
|
with self.mutex:
|
||||||
|
if tpath not in self.busy:
|
||||||
|
break
|
||||||
|
|
||||||
|
with cond:
|
||||||
|
cond.wait()
|
||||||
|
|
||||||
|
try:
|
||||||
|
st = os.stat(tpath)
|
||||||
|
if st.st_size:
|
||||||
|
return tpath
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
def worker(self):
|
||||||
|
while not self.stopping:
|
||||||
|
task = self.q.get()
|
||||||
|
if not task:
|
||||||
|
break
|
||||||
|
|
||||||
|
abspath, tpath = task
|
||||||
|
ext = abspath.split(".")[-1].lower()
|
||||||
|
fun = None
|
||||||
|
if not os.path.exists(tpath):
|
||||||
|
if ext in FMT_PIL:
|
||||||
|
fun = self.conv_pil
|
||||||
|
elif ext in FMT_FF:
|
||||||
|
fun = self.conv_ffmpeg
|
||||||
|
|
||||||
|
if fun:
|
||||||
|
try:
|
||||||
|
fun(abspath, tpath)
|
||||||
|
except:
|
||||||
|
msg = "{} failed on {}\n{}"
|
||||||
|
self.log(msg.format(fun.__name__, abspath, min_ex()), 3)
|
||||||
|
with open(tpath, "wb") as _:
|
||||||
|
pass
|
||||||
|
|
||||||
|
with self.mutex:
|
||||||
|
subs = self.busy[tpath]
|
||||||
|
del self.busy[tpath]
|
||||||
|
|
||||||
|
for x in subs:
|
||||||
|
with x:
|
||||||
|
x.notify_all()
|
||||||
|
|
||||||
|
with self.mutex:
|
||||||
|
self.nthr -= 1
|
||||||
|
|
||||||
|
def conv_pil(self, abspath, tpath):
|
||||||
|
with Image.open(fsenc(abspath)) as im:
|
||||||
|
crop = not self.args.th_no_crop
|
||||||
|
res2 = self.res
|
||||||
|
if crop:
|
||||||
|
res2 = (res2[0] * 2, res2[1] * 2)
|
||||||
|
|
||||||
|
try:
|
||||||
|
im.thumbnail(res2, resample=Image.LANCZOS)
|
||||||
|
if crop:
|
||||||
|
iw, ih = im.size
|
||||||
|
dw, dh = self.res
|
||||||
|
res = (min(iw, dw), min(ih, dh))
|
||||||
|
im = ImageOps.fit(im, res, method=Image.LANCZOS)
|
||||||
|
except:
|
||||||
|
im.thumbnail(self.res)
|
||||||
|
|
||||||
|
fmts = ["RGB", "L"]
|
||||||
|
args = {"quality": 40}
|
||||||
|
|
||||||
|
if tpath.endswith(".webp"):
|
||||||
|
# quality 80 = pillow-default
|
||||||
|
# quality 75 = ffmpeg-default
|
||||||
|
# method 0 = pillow-default, fast
|
||||||
|
# method 4 = ffmpeg-default
|
||||||
|
# method 6 = max, slow
|
||||||
|
fmts += ["RGBA", "LA"]
|
||||||
|
args["method"] = 6
|
||||||
|
else:
|
||||||
|
pass # default q = 75
|
||||||
|
|
||||||
|
if im.mode not in fmts:
|
||||||
|
print("conv {}".format(im.mode))
|
||||||
|
im = im.convert("RGB")
|
||||||
|
|
||||||
|
im.save(tpath, quality=40, method=6)
|
||||||
|
|
||||||
|
def conv_ffmpeg(self, abspath, tpath):
|
||||||
|
ret, _ = ffprobe(abspath)
|
||||||
|
|
||||||
|
ext = abspath.rsplit(".")[-1]
|
||||||
|
if ext in ["h264", "h265"]:
|
||||||
|
seek = []
|
||||||
|
else:
|
||||||
|
dur = ret[".dur"][1] if ".dur" in ret else 4
|
||||||
|
seek = "{:.0f}".format(dur / 3)
|
||||||
|
seek = [b"-ss", seek.encode("utf-8")]
|
||||||
|
|
||||||
|
scale = "scale={0}:{1}:force_original_aspect_ratio="
|
||||||
|
if self.args.th_no_crop:
|
||||||
|
scale += "decrease,setsar=1:1"
|
||||||
|
else:
|
||||||
|
scale += "increase,crop={0}:{1},setsar=1:1"
|
||||||
|
|
||||||
|
scale = scale.format(*list(self.res)).encode("utf-8")
|
||||||
|
# fmt: off
|
||||||
|
cmd = [
|
||||||
|
b"ffmpeg",
|
||||||
|
b"-nostdin",
|
||||||
|
b"-v", b"error",
|
||||||
|
b"-hide_banner"
|
||||||
|
]
|
||||||
|
cmd += seek
|
||||||
|
cmd += [
|
||||||
|
b"-i", fsenc(abspath),
|
||||||
|
b"-vf", scale,
|
||||||
|
b"-vframes", b"1",
|
||||||
|
]
|
||||||
|
# fmt: on
|
||||||
|
|
||||||
|
if tpath.endswith(".jpg"):
|
||||||
|
cmd += [
|
||||||
|
b"-q:v",
|
||||||
|
b"6", # default=??
|
||||||
|
]
|
||||||
|
else:
|
||||||
|
cmd += [
|
||||||
|
b"-q:v",
|
||||||
|
b"50", # default=75
|
||||||
|
b"-compression_level:v",
|
||||||
|
b"6", # default=4, 0=fast, 6=max
|
||||||
|
]
|
||||||
|
|
||||||
|
cmd += [fsenc(tpath)]
|
||||||
|
|
||||||
|
ret, sout, serr = runcmd(*cmd)
|
||||||
|
if ret != 0:
|
||||||
|
msg = ["ff: {}".format(x) for x in serr.split("\n")]
|
||||||
|
self.log("FFmpeg failed:\n" + "\n".join(msg), c="1;30")
|
||||||
|
raise sp.CalledProcessError(ret, (cmd[0], b"...", cmd[-1]))
|
||||||
|
|
||||||
|
def poke(self, tdir):
|
||||||
|
if not self.poke_cd.poke(tdir):
|
||||||
|
return
|
||||||
|
|
||||||
|
ts = int(time.time())
|
||||||
|
try:
|
||||||
|
p1 = os.path.dirname(tdir)
|
||||||
|
p2 = os.path.dirname(p1)
|
||||||
|
for dp in [tdir, p1, p2]:
|
||||||
|
os.utime(fsenc(dp), (ts, ts))
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
def cleaner(self):
|
||||||
|
interval = self.args.th_clean
|
||||||
|
while True:
|
||||||
|
time.sleep(interval)
|
||||||
|
for vol, histpath in self.asrv.vfs.histtab.items():
|
||||||
|
if histpath.startswith(vol):
|
||||||
|
self.log("\033[Jcln {}/\033[A".format(histpath))
|
||||||
|
else:
|
||||||
|
self.log("\033[Jcln {} ({})/\033[A".format(histpath, vol))
|
||||||
|
|
||||||
|
self.clean(histpath)
|
||||||
|
|
||||||
|
self.log("\033[Jcln ok")
|
||||||
|
|
||||||
|
def clean(self, histpath):
|
||||||
|
# self.log("cln {}".format(histpath))
|
||||||
|
maxage = self.args.th_maxage
|
||||||
|
now = time.time()
|
||||||
|
prev_b64 = None
|
||||||
|
prev_fp = None
|
||||||
|
try:
|
||||||
|
ents = os.listdir(histpath)
|
||||||
|
except:
|
||||||
|
return
|
||||||
|
|
||||||
|
for f in sorted(ents):
|
||||||
|
fp = os.path.join(histpath, f)
|
||||||
|
cmp = fp.lower().replace("\\", "/")
|
||||||
|
|
||||||
|
# "top" or b64 prefix/full (a folder)
|
||||||
|
if len(f) <= 3 or len(f) == 24:
|
||||||
|
age = now - os.path.getmtime(fp)
|
||||||
|
if age > maxage:
|
||||||
|
with self.mutex:
|
||||||
|
safe = True
|
||||||
|
for k in self.busy.keys():
|
||||||
|
if k.lower().replace("\\", "/").startswith(cmp):
|
||||||
|
safe = False
|
||||||
|
break
|
||||||
|
|
||||||
|
if safe:
|
||||||
|
self.log("rm -rf [{}]".format(fp))
|
||||||
|
shutil.rmtree(fp, ignore_errors=True)
|
||||||
|
else:
|
||||||
|
self.clean(fp)
|
||||||
|
continue
|
||||||
|
|
||||||
|
# thumb file
|
||||||
|
try:
|
||||||
|
b64, ts, ext = f.split(".")
|
||||||
|
if len(b64) != 24 or len(ts) != 8 or ext not in ["jpg", "webp"]:
|
||||||
|
raise Exception()
|
||||||
|
|
||||||
|
ts = int(ts, 16)
|
||||||
|
except:
|
||||||
|
if f != "dir.txt":
|
||||||
|
self.log("foreign file in thumbs dir: [{}]".format(fp), 1)
|
||||||
|
|
||||||
|
continue
|
||||||
|
|
||||||
|
if b64 == prev_b64:
|
||||||
|
self.log("rm replaced [{}]".format(fp))
|
||||||
|
os.unlink(prev_fp)
|
||||||
|
|
||||||
|
prev_b64 = b64
|
||||||
|
prev_fp = fp
|
||||||
@@ -7,7 +7,7 @@ import time
|
|||||||
import threading
|
import threading
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
from .util import u8safe, s3dec, html_escape, Pebkac
|
from .util import s3dec, Pebkac, min_ex
|
||||||
from .up2k import up2k_wark_from_hashlist
|
from .up2k import up2k_wark_from_hashlist
|
||||||
|
|
||||||
|
|
||||||
@@ -19,10 +19,11 @@ except:
|
|||||||
|
|
||||||
|
|
||||||
class U2idx(object):
|
class U2idx(object):
|
||||||
def __init__(self, args, log_func):
|
def __init__(self, conn):
|
||||||
self.args = args
|
self.log_func = conn.log_func
|
||||||
self.log_func = log_func
|
self.asrv = conn.asrv
|
||||||
self.timeout = args.srch_time
|
self.args = conn.args
|
||||||
|
self.timeout = self.args.srch_time
|
||||||
|
|
||||||
if not HAVE_SQLITE3:
|
if not HAVE_SQLITE3:
|
||||||
self.log("could not load sqlite3; searchign wqill be disabled")
|
self.log("could not load sqlite3; searchign wqill be disabled")
|
||||||
@@ -47,57 +48,143 @@ class U2idx(object):
|
|||||||
fhash = body["hash"]
|
fhash = body["hash"]
|
||||||
wark = up2k_wark_from_hashlist(self.args.salt, fsize, fhash)
|
wark = up2k_wark_from_hashlist(self.args.salt, fsize, fhash)
|
||||||
|
|
||||||
uq = "substr(w,1,16) = ? and w = ?"
|
uq = "where substr(w,1,16) = ? and w = ?"
|
||||||
uv = [wark[:16], wark]
|
uv = [wark[:16], wark]
|
||||||
|
|
||||||
try:
|
try:
|
||||||
return self.run_query(vols, uq, uv, {})[0]
|
return self.run_query(vols, uq, uv)[0]
|
||||||
except Exception as ex:
|
except:
|
||||||
raise Pebkac(500, repr(ex))
|
raise Pebkac(500, min_ex())
|
||||||
|
|
||||||
def get_cur(self, ptop):
|
def get_cur(self, ptop):
|
||||||
cur = self.cur.get(ptop)
|
cur = self.cur.get(ptop)
|
||||||
if cur:
|
if cur:
|
||||||
return cur
|
return cur
|
||||||
|
|
||||||
cur = _open(ptop)
|
histpath = self.asrv.vfs.histtab[ptop]
|
||||||
if not cur:
|
db_path = os.path.join(histpath, "up2k.db")
|
||||||
|
if not os.path.exists(db_path):
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
cur = sqlite3.connect(db_path).cursor()
|
||||||
self.cur[ptop] = cur
|
self.cur[ptop] = cur
|
||||||
return cur
|
return cur
|
||||||
|
|
||||||
def search(self, vols, body):
|
def search(self, vols, uq):
|
||||||
"""search by query params"""
|
"""search by query params"""
|
||||||
if not HAVE_SQLITE3:
|
if not HAVE_SQLITE3:
|
||||||
return []
|
return []
|
||||||
|
|
||||||
qobj = {}
|
q = ""
|
||||||
_conv_sz(qobj, body, "sz_min", "up.sz >= ?")
|
va = []
|
||||||
_conv_sz(qobj, body, "sz_max", "up.sz <= ?")
|
joins = ""
|
||||||
_conv_dt(qobj, body, "dt_min", "up.mt >= ?")
|
is_key = True
|
||||||
_conv_dt(qobj, body, "dt_max", "up.mt <= ?")
|
is_size = False
|
||||||
for seg, dk in [["path", "up.rd"], ["name", "up.fn"]]:
|
is_date = False
|
||||||
if seg in body:
|
kw_key = ["(", ")", "and ", "or ", "not "]
|
||||||
_conv_txt(qobj, body, seg, dk)
|
kw_val = ["==", "=", "!=", ">", ">=", "<", "<=", "like "]
|
||||||
|
ptn_mt = re.compile(r"^\.?[a-z]+$")
|
||||||
|
mt_ctr = 0
|
||||||
|
mt_keycmp = "substr(up.w,1,16)"
|
||||||
|
mt_keycmp2 = None
|
||||||
|
|
||||||
uq, uv = _sqlize(qobj)
|
while True:
|
||||||
|
uq = uq.strip()
|
||||||
|
if not uq:
|
||||||
|
break
|
||||||
|
|
||||||
qobj = {}
|
ok = False
|
||||||
if "tags" in body:
|
for kw in kw_key + kw_val:
|
||||||
_conv_txt(qobj, body, "tags", "mt.v")
|
if uq.startswith(kw):
|
||||||
|
is_key = kw in kw_key
|
||||||
|
uq = uq[len(kw) :]
|
||||||
|
ok = True
|
||||||
|
q += kw
|
||||||
|
break
|
||||||
|
|
||||||
if "adv" in body:
|
if ok:
|
||||||
_conv_adv(qobj, body, "adv")
|
continue
|
||||||
|
|
||||||
|
v, uq = (uq + " ").split(" ", 1)
|
||||||
|
if is_key:
|
||||||
|
is_key = False
|
||||||
|
|
||||||
|
if v == "size":
|
||||||
|
v = "up.sz"
|
||||||
|
is_size = True
|
||||||
|
|
||||||
|
elif v == "date":
|
||||||
|
v = "up.mt"
|
||||||
|
is_date = True
|
||||||
|
|
||||||
|
elif v == "path":
|
||||||
|
v = "up.rd"
|
||||||
|
|
||||||
|
elif v == "name":
|
||||||
|
v = "up.fn"
|
||||||
|
|
||||||
|
elif v == "tags" or ptn_mt.match(v):
|
||||||
|
mt_ctr += 1
|
||||||
|
mt_keycmp2 = "mt{}.w".format(mt_ctr)
|
||||||
|
joins += "inner join mt mt{} on {} = {} ".format(
|
||||||
|
mt_ctr, mt_keycmp, mt_keycmp2
|
||||||
|
)
|
||||||
|
mt_keycmp = mt_keycmp2
|
||||||
|
if v == "tags":
|
||||||
|
v = "mt{0}.v".format(mt_ctr)
|
||||||
|
else:
|
||||||
|
v = "+mt{0}.k = '{1}' and mt{0}.v".format(mt_ctr, v)
|
||||||
|
|
||||||
|
else:
|
||||||
|
raise Pebkac(400, "invalid key [" + v + "]")
|
||||||
|
|
||||||
|
q += v + " "
|
||||||
|
continue
|
||||||
|
|
||||||
|
head = ""
|
||||||
|
tail = ""
|
||||||
|
|
||||||
|
if is_date:
|
||||||
|
is_date = False
|
||||||
|
v = v.upper().rstrip("Z").replace(",", " ").replace("T", " ")
|
||||||
|
while " " in v:
|
||||||
|
v = v.replace(" ", " ")
|
||||||
|
|
||||||
|
for fmt in [
|
||||||
|
"%Y-%m-%d %H:%M:%S",
|
||||||
|
"%Y-%m-%d %H:%M",
|
||||||
|
"%Y-%m-%d %H",
|
||||||
|
"%Y-%m-%d",
|
||||||
|
]:
|
||||||
|
try:
|
||||||
|
v = datetime.strptime(v, fmt).timestamp()
|
||||||
|
break
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
elif is_size:
|
||||||
|
is_size = False
|
||||||
|
v = int(float(v) * 1024 * 1024)
|
||||||
|
|
||||||
|
else:
|
||||||
|
if v.startswith("*"):
|
||||||
|
head = "'%'||"
|
||||||
|
v = v[1:]
|
||||||
|
|
||||||
|
if v.endswith("*"):
|
||||||
|
tail = "||'%'"
|
||||||
|
v = v[:-1]
|
||||||
|
|
||||||
|
q += " {}?{} ".format(head, tail)
|
||||||
|
va.append(v)
|
||||||
|
is_key = True
|
||||||
|
|
||||||
try:
|
try:
|
||||||
return self.run_query(vols, uq, uv, qobj)
|
return self.run_query(vols, joins + "where " + q, va)
|
||||||
except Exception as ex:
|
except Exception as ex:
|
||||||
raise Pebkac(500, repr(ex))
|
raise Pebkac(500, repr(ex))
|
||||||
|
|
||||||
def run_query(self, vols, uq, uv, targs):
|
def run_query(self, vols, uq, uv):
|
||||||
self.log("qs: {} {} , {}".format(uq, repr(uv), repr(targs)))
|
|
||||||
|
|
||||||
done_flag = []
|
done_flag = []
|
||||||
self.active_id = "{:.6f}_{}".format(
|
self.active_id = "{:.6f}_{}".format(
|
||||||
time.time(), threading.current_thread().ident
|
time.time(), threading.current_thread().ident
|
||||||
@@ -108,39 +195,19 @@ class U2idx(object):
|
|||||||
self.active_id,
|
self.active_id,
|
||||||
done_flag,
|
done_flag,
|
||||||
),
|
),
|
||||||
|
name="u2idx-terminator",
|
||||||
)
|
)
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
if not targs:
|
if not uq or not uv:
|
||||||
if not uq:
|
|
||||||
q = "select * from up"
|
q = "select * from up"
|
||||||
v = ()
|
v = ()
|
||||||
else:
|
else:
|
||||||
q = "select * from up where " + uq
|
q = "select up.* from up " + uq
|
||||||
v = tuple(uv)
|
v = tuple(uv)
|
||||||
else:
|
|
||||||
q = "select up.* from up"
|
|
||||||
keycmp = "substr(up.w,1,16)"
|
|
||||||
where = []
|
|
||||||
v = []
|
|
||||||
ctr = 0
|
|
||||||
for tq, tv in sorted(targs.items()):
|
|
||||||
ctr += 1
|
|
||||||
tq = tq.split("\n")[0]
|
|
||||||
keycmp2 = "mt{}.w".format(ctr)
|
|
||||||
q += " inner join mt mt{} on {} = {}".format(ctr, keycmp, keycmp2)
|
|
||||||
keycmp = keycmp2
|
|
||||||
where.append(tq.replace("mt.", keycmp[:-1]))
|
|
||||||
v.append(tv)
|
|
||||||
|
|
||||||
if uq:
|
self.log("qs: {!r} {!r}".format(q, v))
|
||||||
where.append(uq)
|
|
||||||
v.extend(uv)
|
|
||||||
|
|
||||||
q += " where " + (" and ".join(where))
|
|
||||||
|
|
||||||
# self.log("q2: {} {}".format(q, repr(v)))
|
|
||||||
|
|
||||||
ret = []
|
ret = []
|
||||||
lim = 1000
|
lim = 1000
|
||||||
@@ -163,7 +230,7 @@ class U2idx(object):
|
|||||||
if rd.startswith("//") or fn.startswith("//"):
|
if rd.startswith("//") or fn.startswith("//"):
|
||||||
rd, fn = s3dec(rd, fn)
|
rd, fn = s3dec(rd, fn)
|
||||||
|
|
||||||
rp = os.path.join(vtop, rd, fn).replace("\\", "/")
|
rp = "/".join([x for x in [vtop, rd, fn] if x])
|
||||||
sret.append({"ts": int(ts), "sz": sz, "rp": rp, "w": w[:16]})
|
sret.append({"ts": int(ts), "sz": sz, "rp": rp, "w": w[:16]})
|
||||||
|
|
||||||
for hit in sret:
|
for hit in sret:
|
||||||
@@ -178,6 +245,7 @@ class U2idx(object):
|
|||||||
hit["tags"] = tags
|
hit["tags"] = tags
|
||||||
|
|
||||||
ret.extend(sret)
|
ret.extend(sret)
|
||||||
|
# print("[{}] {}".format(ptop, sret))
|
||||||
|
|
||||||
done_flag.append(True)
|
done_flag.append(True)
|
||||||
self.active_id = None
|
self.active_id = None
|
||||||
@@ -198,84 +266,3 @@ class U2idx(object):
|
|||||||
|
|
||||||
if identifier == self.active_id:
|
if identifier == self.active_id:
|
||||||
self.active_cur.connection.interrupt()
|
self.active_cur.connection.interrupt()
|
||||||
|
|
||||||
|
|
||||||
def _open(ptop):
|
|
||||||
db_path = os.path.join(ptop, ".hist", "up2k.db")
|
|
||||||
if os.path.exists(db_path):
|
|
||||||
return sqlite3.connect(db_path).cursor()
|
|
||||||
|
|
||||||
|
|
||||||
def _conv_sz(q, body, k, sql):
|
|
||||||
if k in body:
|
|
||||||
q[sql] = int(float(body[k]) * 1024 * 1024)
|
|
||||||
|
|
||||||
|
|
||||||
def _conv_dt(q, body, k, sql):
|
|
||||||
if k not in body:
|
|
||||||
return
|
|
||||||
|
|
||||||
v = body[k].upper().rstrip("Z").replace(",", " ").replace("T", " ")
|
|
||||||
while " " in v:
|
|
||||||
v = v.replace(" ", " ")
|
|
||||||
|
|
||||||
for fmt in ["%Y-%m-%d %H:%M:%S", "%Y-%m-%d %H:%M", "%Y-%m-%d %H", "%Y-%m-%d"]:
|
|
||||||
try:
|
|
||||||
ts = datetime.strptime(v, fmt).timestamp()
|
|
||||||
break
|
|
||||||
except:
|
|
||||||
ts = None
|
|
||||||
|
|
||||||
if ts:
|
|
||||||
q[sql] = ts
|
|
||||||
|
|
||||||
|
|
||||||
def _conv_txt(q, body, k, sql):
|
|
||||||
for v in body[k].split(" "):
|
|
||||||
inv = ""
|
|
||||||
if v.startswith("-"):
|
|
||||||
inv = "not"
|
|
||||||
v = v[1:]
|
|
||||||
|
|
||||||
if not v:
|
|
||||||
continue
|
|
||||||
|
|
||||||
head = "'%'||"
|
|
||||||
if v.startswith("^"):
|
|
||||||
head = ""
|
|
||||||
v = v[1:]
|
|
||||||
|
|
||||||
tail = "||'%'"
|
|
||||||
if v.endswith("$"):
|
|
||||||
tail = ""
|
|
||||||
v = v[:-1]
|
|
||||||
|
|
||||||
qk = "{} {} like {}?{}".format(sql, inv, head, tail)
|
|
||||||
q[qk + "\n" + v] = u8safe(v)
|
|
||||||
|
|
||||||
|
|
||||||
def _conv_adv(q, body, k):
|
|
||||||
ptn = re.compile(r"^(\.?[a-z]+) *(==?|!=|<=?|>=?) *(.*)$")
|
|
||||||
|
|
||||||
parts = body[k].split(" ")
|
|
||||||
parts = [x.strip() for x in parts if x.strip()]
|
|
||||||
|
|
||||||
for part in parts:
|
|
||||||
m = ptn.match(part)
|
|
||||||
if not m:
|
|
||||||
p = html_escape(part)
|
|
||||||
raise Pebkac(400, "invalid argument [" + p + "]")
|
|
||||||
|
|
||||||
k, op, v = m.groups()
|
|
||||||
qk = "mt.k = '{}' and mt.v {} ?".format(k, op)
|
|
||||||
q[qk + "\n" + v] = u8safe(v)
|
|
||||||
|
|
||||||
|
|
||||||
def _sqlize(qobj):
|
|
||||||
keys = []
|
|
||||||
values = []
|
|
||||||
for k, v in sorted(qobj.items()):
|
|
||||||
keys.append(k.split("\n")[0])
|
|
||||||
values.append(v)
|
|
||||||
|
|
||||||
return " and ".join(keys), values
|
|
||||||
|
|||||||
@@ -31,8 +31,7 @@ from .util import (
|
|||||||
statdir,
|
statdir,
|
||||||
s2hms,
|
s2hms,
|
||||||
)
|
)
|
||||||
from .mtag import MTag
|
from .mtag import MTag, MParser
|
||||||
from .authsrv import AuthSrv
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
HAVE_SQLITE3 = True
|
HAVE_SQLITE3 = True
|
||||||
@@ -49,24 +48,29 @@ class Up2k(object):
|
|||||||
* ~/.config flatfiles for active jobs
|
* ~/.config flatfiles for active jobs
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, broker):
|
def __init__(self, hub):
|
||||||
self.broker = broker
|
self.hub = hub
|
||||||
self.args = broker.args
|
self.asrv = hub.asrv
|
||||||
self.log_func = broker.log
|
self.args = hub.args
|
||||||
|
self.log_func = hub.log
|
||||||
|
|
||||||
# config
|
# config
|
||||||
self.salt = broker.args.salt
|
self.salt = self.args.salt
|
||||||
|
|
||||||
# state
|
# state
|
||||||
self.mutex = threading.Lock()
|
self.mutex = threading.Lock()
|
||||||
self.hashq = Queue()
|
self.hashq = Queue()
|
||||||
self.tagq = Queue()
|
self.tagq = Queue()
|
||||||
|
self.n_hashq = 0
|
||||||
|
self.n_tagq = 0
|
||||||
|
self.volstate = {}
|
||||||
self.registry = {}
|
self.registry = {}
|
||||||
self.entags = {}
|
self.entags = {}
|
||||||
self.flags = {}
|
self.flags = {}
|
||||||
self.cur = {}
|
self.cur = {}
|
||||||
self.mtag = None
|
self.mtag = None
|
||||||
self.pending_tags = None
|
self.pending_tags = None
|
||||||
|
self.mtp_parsers = {}
|
||||||
|
|
||||||
self.mem_cur = None
|
self.mem_cur = None
|
||||||
self.sqlite_ver = None
|
self.sqlite_ver = None
|
||||||
@@ -82,7 +86,7 @@ class Up2k(object):
|
|||||||
if ANYWIN:
|
if ANYWIN:
|
||||||
# usually fails to set lastmod too quickly
|
# usually fails to set lastmod too quickly
|
||||||
self.lastmod_q = Queue()
|
self.lastmod_q = Queue()
|
||||||
thr = threading.Thread(target=self._lastmodder)
|
thr = threading.Thread(target=self._lastmodder, name="up2k-lastmod")
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
@@ -92,31 +96,78 @@ class Up2k(object):
|
|||||||
if not HAVE_SQLITE3:
|
if not HAVE_SQLITE3:
|
||||||
self.log("could not initialize sqlite3, will use in-memory registry only")
|
self.log("could not initialize sqlite3, will use in-memory registry only")
|
||||||
|
|
||||||
# this is kinda jank
|
if self.args.no_fastboot:
|
||||||
auth = AuthSrv(self.args, self.log_func, False)
|
self.deferred_init()
|
||||||
have_e2d = self.init_indexes(auth)
|
else:
|
||||||
|
t = threading.Thread(
|
||||||
|
target=self.deferred_init,
|
||||||
|
name="up2k-deferred-init",
|
||||||
|
)
|
||||||
|
t.daemon = True
|
||||||
|
t.start()
|
||||||
|
|
||||||
|
def deferred_init(self):
|
||||||
|
all_vols = self.asrv.vfs.all_vols
|
||||||
|
have_e2d = self.init_indexes(all_vols)
|
||||||
|
|
||||||
if have_e2d:
|
if have_e2d:
|
||||||
thr = threading.Thread(target=self._snapshot)
|
thr = threading.Thread(target=self._snapshot, name="up2k-snapshot")
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
thr = threading.Thread(target=self._hasher)
|
thr = threading.Thread(target=self._hasher, name="up2k-hasher")
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
if self.mtag:
|
if self.mtag:
|
||||||
thr = threading.Thread(target=self._tagger)
|
thr = threading.Thread(target=self._tagger, name="up2k-tagger")
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
thr = threading.Thread(target=self._run_all_mtp)
|
thr = threading.Thread(target=self._run_all_mtp, name="up2k-mtp-init")
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
def log(self, msg, c=0):
|
def log(self, msg, c=0):
|
||||||
self.log_func("up2k", msg + "\033[K", c)
|
self.log_func("up2k", msg + "\033[K", c)
|
||||||
|
|
||||||
|
def get_state(self):
|
||||||
|
mtpq = 0
|
||||||
|
q = "select count(w) from mt where k = 't:mtp'"
|
||||||
|
got_lock = self.mutex.acquire(timeout=0.5)
|
||||||
|
if got_lock:
|
||||||
|
for cur in self.cur.values():
|
||||||
|
try:
|
||||||
|
mtpq += cur.execute(q).fetchone()[0]
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
self.mutex.release()
|
||||||
|
else:
|
||||||
|
mtpq = "?"
|
||||||
|
|
||||||
|
ret = {
|
||||||
|
"volstate": self.volstate,
|
||||||
|
"scanning": hasattr(self, "pp"),
|
||||||
|
"hashq": self.n_hashq,
|
||||||
|
"tagq": self.n_tagq,
|
||||||
|
"mtpq": mtpq,
|
||||||
|
}
|
||||||
|
return json.dumps(ret, indent=4)
|
||||||
|
|
||||||
|
def rescan(self, all_vols, scan_vols):
|
||||||
|
if hasattr(self, "pp"):
|
||||||
|
return "cannot initiate; scan is already in progress"
|
||||||
|
|
||||||
|
args = (all_vols, scan_vols)
|
||||||
|
t = threading.Thread(
|
||||||
|
target=self.init_indexes,
|
||||||
|
args=args,
|
||||||
|
name="up2k-rescan-{}".format(scan_vols[0]),
|
||||||
|
)
|
||||||
|
t.daemon = True
|
||||||
|
t.start()
|
||||||
|
return None
|
||||||
|
|
||||||
def _vis_job_progress(self, job):
|
def _vis_job_progress(self, job):
|
||||||
perc = 100 - (len(job["need"]) * 100.0 / len(job["hash"]))
|
perc = 100 - (len(job["need"]) * 100.0 / len(job["hash"]))
|
||||||
path = os.path.join(job["ptop"], job["prel"], job["name"])
|
path = os.path.join(job["ptop"], job["prel"], job["name"])
|
||||||
@@ -139,9 +190,9 @@ class Up2k(object):
|
|||||||
|
|
||||||
return True, ret
|
return True, ret
|
||||||
|
|
||||||
def init_indexes(self, auth):
|
def init_indexes(self, all_vols, scan_vols=[]):
|
||||||
self.pp = ProgressPrinter()
|
self.pp = ProgressPrinter()
|
||||||
vols = auth.vfs.all_vols.values()
|
vols = all_vols.values()
|
||||||
t0 = time.time()
|
t0 = time.time()
|
||||||
have_e2d = False
|
have_e2d = False
|
||||||
|
|
||||||
@@ -158,27 +209,42 @@ class Up2k(object):
|
|||||||
self.log(msg, c=3)
|
self.log(msg, c=3)
|
||||||
|
|
||||||
live_vols = []
|
live_vols = []
|
||||||
|
with self.mutex:
|
||||||
|
# only need to protect register_vpath but all in one go feels right
|
||||||
for vol in vols:
|
for vol in vols:
|
||||||
try:
|
try:
|
||||||
os.listdir(vol.realpath)
|
os.listdir(vol.realpath)
|
||||||
live_vols.append(vol)
|
|
||||||
except:
|
except:
|
||||||
|
self.volstate[vol.vpath] = "OFFLINE (cannot access folder)"
|
||||||
self.log("cannot access " + vol.realpath, c=1)
|
self.log("cannot access " + vol.realpath, c=1)
|
||||||
|
continue
|
||||||
|
|
||||||
|
if scan_vols and vol.vpath not in scan_vols:
|
||||||
|
continue
|
||||||
|
|
||||||
|
if not self.register_vpath(vol.realpath, vol.flags):
|
||||||
|
# self.log("db not enable for {}".format(m, vol.realpath))
|
||||||
|
continue
|
||||||
|
|
||||||
|
live_vols.append(vol)
|
||||||
|
|
||||||
|
if vol.vpath not in self.volstate:
|
||||||
|
self.volstate[vol.vpath] = "OFFLINE (pending initialization)"
|
||||||
|
|
||||||
vols = live_vols
|
vols = live_vols
|
||||||
|
need_vac = {}
|
||||||
|
|
||||||
need_mtag = False
|
need_mtag = False
|
||||||
for vol in vols:
|
for vol in vols:
|
||||||
if "e2t" in vol.flags:
|
if "e2t" in vol.flags:
|
||||||
need_mtag = True
|
need_mtag = True
|
||||||
|
|
||||||
if need_mtag:
|
if need_mtag and not self.mtag:
|
||||||
self.mtag = MTag(self.log_func, self.args)
|
self.mtag = MTag(self.log_func, self.args)
|
||||||
if not self.mtag.usable:
|
if not self.mtag.usable:
|
||||||
self.mtag = None
|
self.mtag = None
|
||||||
|
|
||||||
# e2ds(a) volumes first,
|
# e2ds(a) volumes first
|
||||||
# also covers tags where e2ts is set
|
|
||||||
for vol in vols:
|
for vol in vols:
|
||||||
en = {}
|
en = {}
|
||||||
if "mte" in vol.flags:
|
if "mte" in vol.flags:
|
||||||
@@ -190,26 +256,45 @@ class Up2k(object):
|
|||||||
have_e2d = True
|
have_e2d = True
|
||||||
|
|
||||||
if "e2ds" in vol.flags:
|
if "e2ds" in vol.flags:
|
||||||
r = self._build_file_index(vol, vols)
|
self.volstate[vol.vpath] = "busy (hashing files)"
|
||||||
if not r:
|
_, vac = self._build_file_index(vol, list(all_vols.values()))
|
||||||
needed_mutagen = True
|
if vac:
|
||||||
|
need_vac[vol] = True
|
||||||
|
|
||||||
|
if "e2ts" not in vol.flags:
|
||||||
|
m = "online, idle"
|
||||||
|
else:
|
||||||
|
m = "online (tags pending)"
|
||||||
|
|
||||||
|
self.volstate[vol.vpath] = m
|
||||||
|
|
||||||
# open the rest + do any e2ts(a)
|
# open the rest + do any e2ts(a)
|
||||||
needed_mutagen = False
|
needed_mutagen = False
|
||||||
for vol in vols:
|
for vol in vols:
|
||||||
r = self.register_vpath(vol.realpath, vol.flags)
|
if "e2ts" not in vol.flags:
|
||||||
if not r or "e2ts" not in vol.flags:
|
|
||||||
continue
|
continue
|
||||||
|
|
||||||
cur, db_path, sz0 = r
|
m = "online (reading tags)"
|
||||||
n_add, n_rm, success = self._build_tags_index(vol.realpath)
|
self.volstate[vol.vpath] = m
|
||||||
|
self.log("{} [{}]".format(m, vol.realpath))
|
||||||
|
|
||||||
|
nadd, nrm, success = self._build_tags_index(vol)
|
||||||
if not success:
|
if not success:
|
||||||
needed_mutagen = True
|
needed_mutagen = True
|
||||||
|
|
||||||
if n_add or n_rm:
|
if nadd or nrm:
|
||||||
self.vac(cur, db_path, n_add, n_rm, sz0)
|
need_vac[vol] = True
|
||||||
|
|
||||||
|
self.volstate[vol.vpath] = "online (mtp soon)"
|
||||||
|
|
||||||
|
for vol in need_vac:
|
||||||
|
cur, _ = self.register_vpath(vol.realpath, vol.flags)
|
||||||
|
with self.mutex:
|
||||||
|
cur.connection.commit()
|
||||||
|
cur.execute("vacuum")
|
||||||
|
|
||||||
self.pp.end = True
|
self.pp.end = True
|
||||||
|
|
||||||
msg = "{} volumes in {:.2f} sec"
|
msg = "{} volumes in {:.2f} sec"
|
||||||
self.log(msg.format(len(vols), time.time() - t0))
|
self.log(msg.format(len(vols), time.time() - t0))
|
||||||
|
|
||||||
@@ -217,11 +302,31 @@ class Up2k(object):
|
|||||||
msg = "could not read tags because no backends are available (mutagen or ffprobe)"
|
msg = "could not read tags because no backends are available (mutagen or ffprobe)"
|
||||||
self.log(msg, c=1)
|
self.log(msg, c=1)
|
||||||
|
|
||||||
|
thr = None
|
||||||
|
if self.mtag:
|
||||||
|
m = "online (running mtp)"
|
||||||
|
if scan_vols:
|
||||||
|
thr = threading.Thread(target=self._run_all_mtp, name="up2k-mtp-scan")
|
||||||
|
thr.daemon = True
|
||||||
|
else:
|
||||||
|
del self.pp
|
||||||
|
m = "online, idle"
|
||||||
|
|
||||||
|
for vol in vols:
|
||||||
|
self.volstate[vol.vpath] = m
|
||||||
|
|
||||||
|
if thr:
|
||||||
|
thr.start()
|
||||||
|
|
||||||
return have_e2d
|
return have_e2d
|
||||||
|
|
||||||
def register_vpath(self, ptop, flags):
|
def register_vpath(self, ptop, flags):
|
||||||
with self.mutex:
|
histpath = self.asrv.vfs.histtab[ptop]
|
||||||
|
db_path = os.path.join(histpath, "up2k.db")
|
||||||
if ptop in self.registry:
|
if ptop in self.registry:
|
||||||
|
try:
|
||||||
|
return [self.cur[ptop], db_path]
|
||||||
|
except:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
_, flags = self._expr_idx_filter(flags)
|
_, flags = self._expr_idx_filter(flags)
|
||||||
@@ -237,14 +342,19 @@ class Up2k(object):
|
|||||||
self.log(" ".join(sorted(a)) + "\033[0m")
|
self.log(" ".join(sorted(a)) + "\033[0m")
|
||||||
|
|
||||||
reg = {}
|
reg = {}
|
||||||
path = os.path.join(ptop, ".hist", "up2k.snap")
|
path = os.path.join(histpath, "up2k.snap")
|
||||||
if "e2d" in flags and os.path.exists(path):
|
if "e2d" in flags and os.path.exists(path):
|
||||||
with gzip.GzipFile(path, "rb") as f:
|
with gzip.GzipFile(path, "rb") as f:
|
||||||
j = f.read().decode("utf-8")
|
j = f.read().decode("utf-8")
|
||||||
|
|
||||||
reg = json.loads(j)
|
reg2 = json.loads(j)
|
||||||
for _, job in reg.items():
|
for k, job in reg2.items():
|
||||||
|
path = os.path.join(job["ptop"], job["prel"], job["name"])
|
||||||
|
if os.path.exists(fsenc(path)):
|
||||||
|
reg[k] = job
|
||||||
job["poke"] = time.time()
|
job["poke"] = time.time()
|
||||||
|
else:
|
||||||
|
self.log("ign deleted file in snap: [{}]".format(path))
|
||||||
|
|
||||||
m = "loaded snap {} |{}|".format(path, len(reg.keys()))
|
m = "loaded snap {} |{}|".format(path, len(reg.keys()))
|
||||||
m = [m] + self._vis_reg_progress(reg)
|
m = [m] + self._vis_reg_progress(reg)
|
||||||
@@ -256,22 +366,14 @@ class Up2k(object):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
os.mkdir(os.path.join(ptop, ".hist"))
|
os.makedirs(histpath)
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
db_path = os.path.join(ptop, ".hist", "up2k.db")
|
|
||||||
if ptop in self.cur:
|
|
||||||
return None
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
sz0 = 0
|
|
||||||
if os.path.exists(db_path):
|
|
||||||
sz0 = os.path.getsize(db_path) // 1024
|
|
||||||
|
|
||||||
cur = self._open_db(db_path)
|
cur = self._open_db(db_path)
|
||||||
self.cur[ptop] = cur
|
self.cur[ptop] = cur
|
||||||
return [cur, db_path, sz0]
|
return [cur, db_path]
|
||||||
except:
|
except:
|
||||||
msg = "cannot use database at [{}]:\n{}"
|
msg = "cannot use database at [{}]:\n{}"
|
||||||
self.log(msg.format(ptop, traceback.format_exc()))
|
self.log(msg.format(ptop, traceback.format_exc()))
|
||||||
@@ -281,12 +383,11 @@ class Up2k(object):
|
|||||||
def _build_file_index(self, vol, all_vols):
|
def _build_file_index(self, vol, all_vols):
|
||||||
do_vac = False
|
do_vac = False
|
||||||
top = vol.realpath
|
top = vol.realpath
|
||||||
reg = self.register_vpath(top, vol.flags)
|
nohash = "dhash" in vol.flags
|
||||||
if not reg:
|
with self.mutex:
|
||||||
return
|
cur, _ = self.register_vpath(top, vol.flags)
|
||||||
|
|
||||||
_, db_path, sz0 = reg
|
dbw = [cur, 0, time.time()]
|
||||||
dbw = [reg[0], 0, time.time()]
|
|
||||||
self.pp.n = next(dbw[0].execute("select count(w) from up"))[0]
|
self.pp.n = next(dbw[0].execute("select count(w) from up"))[0]
|
||||||
|
|
||||||
excl = [
|
excl = [
|
||||||
@@ -294,46 +395,36 @@ class Up2k(object):
|
|||||||
for d in all_vols
|
for d in all_vols
|
||||||
if d != vol and (d.vpath.startswith(vol.vpath + "/") or not vol.vpath)
|
if d != vol and (d.vpath.startswith(vol.vpath + "/") or not vol.vpath)
|
||||||
]
|
]
|
||||||
n_add = self._build_dir(dbw, top, set(excl), top)
|
if WINDOWS:
|
||||||
|
excl = [x.replace("/", "\\") for x in excl]
|
||||||
|
|
||||||
|
n_add = self._build_dir(dbw, top, set(excl), top, nohash)
|
||||||
n_rm = self._drop_lost(dbw[0], top)
|
n_rm = self._drop_lost(dbw[0], top)
|
||||||
if dbw[1]:
|
if dbw[1]:
|
||||||
self.log("commit {} new files".format(dbw[1]))
|
self.log("commit {} new files".format(dbw[1]))
|
||||||
dbw[0].connection.commit()
|
dbw[0].connection.commit()
|
||||||
|
|
||||||
n_add, n_rm, success = self._build_tags_index(vol.realpath)
|
return True, n_add or n_rm or do_vac
|
||||||
|
|
||||||
dbw[0].connection.commit()
|
def _build_dir(self, dbw, top, excl, cdir, nohash):
|
||||||
if n_add or n_rm or do_vac:
|
|
||||||
self.vac(dbw[0], db_path, n_add, n_rm, sz0)
|
|
||||||
|
|
||||||
return success
|
|
||||||
|
|
||||||
def vac(self, cur, db_path, n_add, n_rm, sz0):
|
|
||||||
sz1 = os.path.getsize(db_path) // 1024
|
|
||||||
cur.execute("vacuum")
|
|
||||||
sz2 = os.path.getsize(db_path) // 1024
|
|
||||||
msg = "{} new, {} del, {} kB vacced, {} kB gain, {} kB now".format(
|
|
||||||
n_add, n_rm, sz1 - sz2, sz2 - sz0, sz2
|
|
||||||
)
|
|
||||||
self.log(msg)
|
|
||||||
|
|
||||||
def _build_dir(self, dbw, top, excl, cdir):
|
|
||||||
self.pp.msg = "a{} {}".format(self.pp.n, cdir)
|
self.pp.msg = "a{} {}".format(self.pp.n, cdir)
|
||||||
histdir = os.path.join(top, ".hist")
|
histpath = self.asrv.vfs.histtab[top]
|
||||||
ret = 0
|
ret = 0
|
||||||
for iname, inf in statdir(self.log, not self.args.no_scandir, False, cdir):
|
g = statdir(self.log, not self.args.no_scandir, False, cdir)
|
||||||
|
for iname, inf in sorted(g):
|
||||||
abspath = os.path.join(cdir, iname)
|
abspath = os.path.join(cdir, iname)
|
||||||
lmod = int(inf.st_mtime)
|
lmod = int(inf.st_mtime)
|
||||||
|
sz = inf.st_size
|
||||||
if stat.S_ISDIR(inf.st_mode):
|
if stat.S_ISDIR(inf.st_mode):
|
||||||
if abspath in excl or abspath == histdir:
|
if abspath in excl or abspath == histpath:
|
||||||
continue
|
continue
|
||||||
# self.log(" dir: {}".format(abspath))
|
# self.log(" dir: {}".format(abspath))
|
||||||
ret += self._build_dir(dbw, top, excl, abspath)
|
ret += self._build_dir(dbw, top, excl, abspath, nohash)
|
||||||
else:
|
else:
|
||||||
# self.log("file: {}".format(abspath))
|
# self.log("file: {}".format(abspath))
|
||||||
rp = abspath[len(top) :].replace("\\", "/").strip("/")
|
rp = abspath[len(top) :].replace("\\", "/").strip("/")
|
||||||
rd, fn = rp.rsplit("/", 1) if "/" in rp else ["", rp]
|
rd, fn = rp.rsplit("/", 1) if "/" in rp else ["", rp]
|
||||||
sql = "select * from up where rd = ? and fn = ?"
|
sql = "select w, mt, sz from up where rd = ? and fn = ?"
|
||||||
try:
|
try:
|
||||||
c = dbw[0].execute(sql, (rd, fn))
|
c = dbw[0].execute(sql, (rd, fn))
|
||||||
except:
|
except:
|
||||||
@@ -342,18 +433,18 @@ class Up2k(object):
|
|||||||
in_db = list(c.fetchall())
|
in_db = list(c.fetchall())
|
||||||
if in_db:
|
if in_db:
|
||||||
self.pp.n -= 1
|
self.pp.n -= 1
|
||||||
_, dts, dsz, _, _ = in_db[0]
|
dw, dts, dsz = in_db[0]
|
||||||
if len(in_db) > 1:
|
if len(in_db) > 1:
|
||||||
m = "WARN: multiple entries: [{}] => [{}] |{}|\n{}"
|
m = "WARN: multiple entries: [{}] => [{}] |{}|\n{}"
|
||||||
rep_db = "\n".join([repr(x) for x in in_db])
|
rep_db = "\n".join([repr(x) for x in in_db])
|
||||||
self.log(m.format(top, rp, len(in_db), rep_db))
|
self.log(m.format(top, rp, len(in_db), rep_db))
|
||||||
dts = -1
|
dts = -1
|
||||||
|
|
||||||
if dts == lmod and dsz == inf.st_size:
|
if dts == lmod and dsz == sz and (nohash or dw[0] != "#"):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
m = "reindex [{}] => [{}] ({}/{}) ({}/{})".format(
|
m = "reindex [{}] => [{}] ({}/{}) ({}/{})".format(
|
||||||
top, rp, dts, lmod, dsz, inf.st_size
|
top, rp, dts, lmod, dsz, sz
|
||||||
)
|
)
|
||||||
self.log(m)
|
self.log(m)
|
||||||
self.db_rm(dbw[0], rd, fn)
|
self.db_rm(dbw[0], rd, fn)
|
||||||
@@ -362,7 +453,11 @@ class Up2k(object):
|
|||||||
in_db = None
|
in_db = None
|
||||||
|
|
||||||
self.pp.msg = "a{} {}".format(self.pp.n, abspath)
|
self.pp.msg = "a{} {}".format(self.pp.n, abspath)
|
||||||
if inf.st_size > 1024 * 1024:
|
|
||||||
|
if nohash:
|
||||||
|
wark = up2k_wark_from_metadata(self.salt, sz, lmod, rd, fn)
|
||||||
|
else:
|
||||||
|
if sz > 1024 * 1024:
|
||||||
self.log("file: {}".format(abspath))
|
self.log("file: {}".format(abspath))
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@@ -371,8 +466,9 @@ class Up2k(object):
|
|||||||
self.log("hash: {} @ [{}]".format(repr(ex), abspath))
|
self.log("hash: {} @ [{}]".format(repr(ex), abspath))
|
||||||
continue
|
continue
|
||||||
|
|
||||||
wark = up2k_wark_from_hashlist(self.salt, inf.st_size, hashes)
|
wark = up2k_wark_from_hashlist(self.salt, sz, hashes)
|
||||||
self.db_add(dbw[0], wark, rd, fn, lmod, inf.st_size)
|
|
||||||
|
self.db_add(dbw[0], wark, rd, fn, lmod, sz)
|
||||||
dbw[1] += 1
|
dbw[1] += 1
|
||||||
ret += 1
|
ret += 1
|
||||||
td = time.time() - dbw[2]
|
td = time.time() - dbw[2]
|
||||||
@@ -410,25 +506,29 @@ class Up2k(object):
|
|||||||
|
|
||||||
return len(rm)
|
return len(rm)
|
||||||
|
|
||||||
def _build_tags_index(self, ptop):
|
def _build_tags_index(self, vol):
|
||||||
|
ptop = vol.realpath
|
||||||
|
with self.mutex:
|
||||||
|
_, db_path = self.register_vpath(ptop, vol.flags)
|
||||||
entags = self.entags[ptop]
|
entags = self.entags[ptop]
|
||||||
flags = self.flags[ptop]
|
flags = self.flags[ptop]
|
||||||
cur = self.cur[ptop]
|
cur = self.cur[ptop]
|
||||||
|
|
||||||
n_add = 0
|
n_add = 0
|
||||||
n_rm = 0
|
n_rm = 0
|
||||||
n_buf = 0
|
n_buf = 0
|
||||||
last_write = time.time()
|
last_write = time.time()
|
||||||
|
|
||||||
if "e2tsr" in flags:
|
if "e2tsr" in flags:
|
||||||
|
with self.mutex:
|
||||||
n_rm = cur.execute("select count(w) from mt").fetchone()[0]
|
n_rm = cur.execute("select count(w) from mt").fetchone()[0]
|
||||||
if n_rm:
|
if n_rm:
|
||||||
self.log("discarding {} media tags for a full rescan".format(n_rm))
|
self.log("discarding {} media tags for a full rescan".format(n_rm))
|
||||||
cur.execute("delete from mt")
|
cur.execute("delete from mt")
|
||||||
else:
|
|
||||||
self.log("volume has e2tsr but there are no media tags to discard")
|
|
||||||
|
|
||||||
# integrity: drop tags for tracks that were deleted
|
# integrity: drop tags for tracks that were deleted
|
||||||
if "e2t" in flags:
|
if "e2t" in flags:
|
||||||
|
with self.mutex:
|
||||||
drops = []
|
drops = []
|
||||||
c2 = cur.connection.cursor()
|
c2 = cur.connection.cursor()
|
||||||
up_q = "select w from up where substr(w,1,16) = ?"
|
up_q = "select w from up where substr(w,1,16) = ?"
|
||||||
@@ -449,6 +549,10 @@ class Up2k(object):
|
|||||||
return n_add, n_rm, True
|
return n_add, n_rm, True
|
||||||
|
|
||||||
# add tags for new files
|
# add tags for new files
|
||||||
|
gcur = cur
|
||||||
|
with self.mutex:
|
||||||
|
gcur.connection.commit()
|
||||||
|
|
||||||
if "e2ts" in flags:
|
if "e2ts" in flags:
|
||||||
if not self.mtag:
|
if not self.mtag:
|
||||||
return n_add, n_rm, False
|
return n_add, n_rm, False
|
||||||
@@ -457,8 +561,10 @@ class Up2k(object):
|
|||||||
if self.mtag.prefer_mt and not self.args.no_mtag_mt:
|
if self.mtag.prefer_mt and not self.args.no_mtag_mt:
|
||||||
mpool = self._start_mpool()
|
mpool = self._start_mpool()
|
||||||
|
|
||||||
c2 = cur.connection.cursor()
|
conn = sqlite3.connect(db_path, timeout=15)
|
||||||
c3 = cur.connection.cursor()
|
cur = conn.cursor()
|
||||||
|
c2 = conn.cursor()
|
||||||
|
c3 = conn.cursor()
|
||||||
n_left = cur.execute("select count(w) from up").fetchone()[0]
|
n_left = cur.execute("select count(w) from up").fetchone()[0]
|
||||||
for w, rd, fn in cur.execute("select w, rd, fn from up"):
|
for w, rd, fn in cur.execute("select w, rd, fn from up"):
|
||||||
n_left -= 1
|
n_left -= 1
|
||||||
@@ -480,6 +586,7 @@ class Up2k(object):
|
|||||||
n_tags = self._tag_file(c3, *args)
|
n_tags = self._tag_file(c3, *args)
|
||||||
else:
|
else:
|
||||||
mpool.put(["mtag"] + args)
|
mpool.put(["mtag"] + args)
|
||||||
|
with self.mutex:
|
||||||
n_tags = len(self._flush_mpool(c3))
|
n_tags = len(self._flush_mpool(c3))
|
||||||
|
|
||||||
n_add += n_tags
|
n_add += n_tags
|
||||||
@@ -492,15 +599,23 @@ class Up2k(object):
|
|||||||
last_write = time.time()
|
last_write = time.time()
|
||||||
n_buf = 0
|
n_buf = 0
|
||||||
|
|
||||||
self._stop_mpool(mpool, c3)
|
if mpool:
|
||||||
|
self._stop_mpool(mpool)
|
||||||
|
with self.mutex:
|
||||||
|
n_add += len(self._flush_mpool(c3))
|
||||||
|
|
||||||
|
conn.commit()
|
||||||
c3.close()
|
c3.close()
|
||||||
c2.close()
|
c2.close()
|
||||||
|
cur.close()
|
||||||
|
conn.close()
|
||||||
|
|
||||||
|
with self.mutex:
|
||||||
|
gcur.connection.commit()
|
||||||
|
|
||||||
return n_add, n_rm, True
|
return n_add, n_rm, True
|
||||||
|
|
||||||
def _flush_mpool(self, wcur):
|
def _flush_mpool(self, wcur):
|
||||||
with self.mutex:
|
|
||||||
ret = []
|
ret = []
|
||||||
for x in self.pending_tags:
|
for x in self.pending_tags:
|
||||||
self._tag_file(wcur, *x)
|
self._tag_file(wcur, *x)
|
||||||
@@ -511,9 +626,6 @@ class Up2k(object):
|
|||||||
|
|
||||||
def _run_all_mtp(self):
|
def _run_all_mtp(self):
|
||||||
t0 = time.time()
|
t0 = time.time()
|
||||||
self.mtp_audio = {}
|
|
||||||
self.mtp_force = {}
|
|
||||||
self.mtp_parsers = {}
|
|
||||||
for ptop, flags in self.flags.items():
|
for ptop, flags in self.flags.items():
|
||||||
if "mtp" in flags:
|
if "mtp" in flags:
|
||||||
self._run_one_mtp(ptop)
|
self._run_one_mtp(ptop)
|
||||||
@@ -522,58 +634,26 @@ class Up2k(object):
|
|||||||
msg = "mtp finished in {:.2f} sec ({})"
|
msg = "mtp finished in {:.2f} sec ({})"
|
||||||
self.log(msg.format(td, s2hms(td, True)))
|
self.log(msg.format(td, s2hms(td, True)))
|
||||||
|
|
||||||
def _run_one_mtp(self, ptop):
|
del self.pp
|
||||||
db_path = os.path.join(ptop, ".hist", "up2k.db")
|
for k in list(self.volstate.keys()):
|
||||||
sz0 = os.path.getsize(db_path) // 1024
|
if "OFFLINE" not in self.volstate[k]:
|
||||||
|
self.volstate[k] = "online, idle"
|
||||||
|
|
||||||
|
def _run_one_mtp(self, ptop):
|
||||||
entags = self.entags[ptop]
|
entags = self.entags[ptop]
|
||||||
|
|
||||||
audio = {} # [r]equire [n]ot [d]ontcare
|
|
||||||
force = {} # bool
|
|
||||||
timeout = {} # int
|
|
||||||
parsers = {}
|
parsers = {}
|
||||||
for parser in self.flags[ptop]["mtp"]:
|
for parser in self.flags[ptop]["mtp"]:
|
||||||
orig = parser
|
|
||||||
tag, parser = parser.split("=", 1)
|
|
||||||
if tag not in entags:
|
|
||||||
continue
|
|
||||||
|
|
||||||
audio[tag] = "y"
|
|
||||||
|
|
||||||
while True:
|
|
||||||
try:
|
try:
|
||||||
bp = os.path.expanduser(parser)
|
parser = MParser(parser)
|
||||||
if os.path.exists(bp):
|
|
||||||
parsers[tag] = [bp, timeout.get(tag, 30)]
|
|
||||||
break
|
|
||||||
except:
|
except:
|
||||||
pass
|
self.log("invalid argument: " + parser, 1)
|
||||||
|
|
||||||
try:
|
|
||||||
arg, parser = parser.split(",", 1)
|
|
||||||
arg = arg.lower()
|
|
||||||
|
|
||||||
if arg.startswith("a"):
|
|
||||||
audio[tag] = arg[1:]
|
|
||||||
continue
|
|
||||||
|
|
||||||
if arg == "f":
|
|
||||||
force[tag] = True
|
|
||||||
continue
|
|
||||||
|
|
||||||
if arg.startswith("t"):
|
|
||||||
timeout[tag] = int(arg[1:])
|
|
||||||
continue
|
|
||||||
|
|
||||||
raise Exception()
|
|
||||||
|
|
||||||
except:
|
|
||||||
self.log("invalid argument: " + orig, 1)
|
|
||||||
return
|
return
|
||||||
|
|
||||||
# todo audio/force => parser attributes
|
for tag in entags:
|
||||||
self.mtp_audio[ptop] = audio
|
if tag in parser.tags:
|
||||||
self.mtp_force[ptop] = force
|
parsers[parser.tag] = parser
|
||||||
|
|
||||||
self.mtp_parsers[ptop] = parsers
|
self.mtp_parsers[ptop] = parsers
|
||||||
|
|
||||||
q = "select count(w) from mt where k = 't:mtp'"
|
q = "select count(w) from mt where k = 't:mtp'"
|
||||||
@@ -606,7 +686,7 @@ class Up2k(object):
|
|||||||
have = cur.execute(q, (w,)).fetchall()
|
have = cur.execute(q, (w,)).fetchall()
|
||||||
have = [x[0] for x in have]
|
have = [x[0] for x in have]
|
||||||
|
|
||||||
parsers = self._get_parsers(ptop, have)
|
parsers = self._get_parsers(ptop, have, abspath)
|
||||||
if not parsers:
|
if not parsers:
|
||||||
to_delete[w] = True
|
to_delete[w] = True
|
||||||
n_left -= 1
|
n_left -= 1
|
||||||
@@ -618,9 +698,8 @@ class Up2k(object):
|
|||||||
jobs.append([parsers, None, w, abspath])
|
jobs.append([parsers, None, w, abspath])
|
||||||
in_progress[w] = True
|
in_progress[w] = True
|
||||||
|
|
||||||
done = self._flush_mpool(wcur)
|
|
||||||
|
|
||||||
with self.mutex:
|
with self.mutex:
|
||||||
|
done = self._flush_mpool(wcur)
|
||||||
for w in done:
|
for w in done:
|
||||||
to_delete[w] = True
|
to_delete[w] = True
|
||||||
in_progress.pop(w)
|
in_progress.pop(w)
|
||||||
@@ -661,56 +740,75 @@ class Up2k(object):
|
|||||||
with self.mutex:
|
with self.mutex:
|
||||||
cur.connection.commit()
|
cur.connection.commit()
|
||||||
|
|
||||||
done = self._stop_mpool(mpool, wcur)
|
self._stop_mpool(mpool)
|
||||||
with self.mutex:
|
with self.mutex:
|
||||||
|
done = self._flush_mpool(wcur)
|
||||||
for w in done:
|
for w in done:
|
||||||
q = "delete from mt where w = ? and k = 't:mtp'"
|
q = "delete from mt where w = ? and k = 't:mtp'"
|
||||||
cur.execute(q, (w,))
|
cur.execute(q, (w,))
|
||||||
|
|
||||||
cur.connection.commit()
|
cur.connection.commit()
|
||||||
if n_done:
|
if n_done:
|
||||||
self.vac(cur, db_path, n_done, 0, sz0)
|
cur.execute("vacuum")
|
||||||
|
|
||||||
wcur.close()
|
wcur.close()
|
||||||
cur.close()
|
cur.close()
|
||||||
|
|
||||||
def _get_parsers(self, ptop, have):
|
def _get_parsers(self, ptop, have, abspath):
|
||||||
audio = self.mtp_audio[ptop]
|
try:
|
||||||
force = self.mtp_force[ptop]
|
all_parsers = self.mtp_parsers[ptop]
|
||||||
|
except:
|
||||||
|
return {}
|
||||||
|
|
||||||
entags = self.entags[ptop]
|
entags = self.entags[ptop]
|
||||||
parsers = {}
|
parsers = {}
|
||||||
for k, v in self.mtp_parsers[ptop].items():
|
for k, v in all_parsers.items():
|
||||||
if ".dur" in entags:
|
if "ac" in entags or ".aq" in entags:
|
||||||
if ".dur" in have:
|
if "ac" in have or ".aq" in have:
|
||||||
# is audio, require non-audio?
|
# is audio, require non-audio?
|
||||||
if audio[k] == "n":
|
if v.audio == "n":
|
||||||
continue
|
continue
|
||||||
# is not audio, require audio?
|
# is not audio, require audio?
|
||||||
elif audio[k] == "y":
|
elif v.audio == "y":
|
||||||
|
continue
|
||||||
|
|
||||||
|
if v.ext:
|
||||||
|
match = False
|
||||||
|
for ext in v.ext:
|
||||||
|
if abspath.lower().endswith("." + ext):
|
||||||
|
match = True
|
||||||
|
break
|
||||||
|
|
||||||
|
if not match:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
parsers[k] = v
|
parsers[k] = v
|
||||||
|
|
||||||
parsers = {k: v for k, v in parsers.items() if k in force or k not in have}
|
parsers = {k: v for k, v in parsers.items() if v.force or k not in have}
|
||||||
return parsers
|
return parsers
|
||||||
|
|
||||||
def _start_mpool(self):
|
def _start_mpool(self):
|
||||||
# mp.pool.ThreadPool and concurrent.futures.ThreadPoolExecutor
|
# mp.pool.ThreadPool and concurrent.futures.ThreadPoolExecutor
|
||||||
# both do crazy runahead so lets reinvent another wheel
|
# both do crazy runahead so lets reinvent another wheel
|
||||||
nw = os.cpu_count() if hasattr(os, "cpu_count") else 4
|
nw = os.cpu_count() if hasattr(os, "cpu_count") else 4
|
||||||
|
if self.args.no_mtag_mt:
|
||||||
|
nw = 1
|
||||||
|
|
||||||
if self.pending_tags is None:
|
if self.pending_tags is None:
|
||||||
self.log("using {}x {}".format(nw, self.mtag.backend))
|
self.log("using {}x {}".format(nw, self.mtag.backend))
|
||||||
self.pending_tags = []
|
self.pending_tags = []
|
||||||
|
|
||||||
mpool = Queue(nw)
|
mpool = Queue(nw)
|
||||||
for _ in range(nw):
|
for _ in range(nw):
|
||||||
thr = threading.Thread(target=self._tag_thr, args=(mpool,))
|
thr = threading.Thread(
|
||||||
|
target=self._tag_thr, args=(mpool,), name="up2k-mpool"
|
||||||
|
)
|
||||||
thr.daemon = True
|
thr.daemon = True
|
||||||
thr.start()
|
thr.start()
|
||||||
|
|
||||||
return mpool
|
return mpool
|
||||||
|
|
||||||
def _stop_mpool(self, mpool, wcur):
|
def _stop_mpool(self, mpool):
|
||||||
if not mpool:
|
if not mpool:
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -718,8 +816,6 @@ class Up2k(object):
|
|||||||
mpool.put(None)
|
mpool.put(None)
|
||||||
|
|
||||||
mpool.join()
|
mpool.join()
|
||||||
done = self._flush_mpool(wcur)
|
|
||||||
return done
|
|
||||||
|
|
||||||
def _tag_thr(self, q):
|
def _tag_thr(self, q):
|
||||||
while True:
|
while True:
|
||||||
@@ -737,6 +833,7 @@ class Up2k(object):
|
|||||||
vtags = [
|
vtags = [
|
||||||
"\033[36m{} \033[33m{}".format(k, v) for k, v in tags.items()
|
"\033[36m{} \033[33m{}".format(k, v) for k, v in tags.items()
|
||||||
]
|
]
|
||||||
|
if vtags:
|
||||||
self.log("{}\033[0m [{}]".format(" ".join(vtags), abspath))
|
self.log("{}\033[0m [{}]".format(" ".join(vtags), abspath))
|
||||||
|
|
||||||
with self.mutex:
|
with self.mutex:
|
||||||
@@ -866,7 +963,7 @@ class Up2k(object):
|
|||||||
def _create_v3(self, cur):
|
def _create_v3(self, cur):
|
||||||
"""
|
"""
|
||||||
collision in 2^(n/2) files where n = bits (6 bits/ch)
|
collision in 2^(n/2) files where n = bits (6 bits/ch)
|
||||||
10*6/2 = 2^30 = 1'073'741'824, 24.1mb idx
|
10*6/2 = 2^30 = 1'073'741'824, 24.1mb idx 1<<(3*10)
|
||||||
12*6/2 = 2^36 = 68'719'476'736, 24.8mb idx
|
12*6/2 = 2^36 = 68'719'476'736, 24.8mb idx
|
||||||
16*6/2 = 2^48 = 281'474'976'710'656, 26.1mb idx
|
16*6/2 = 2^48 = 281'474'976'710'656, 26.1mb idx
|
||||||
"""
|
"""
|
||||||
@@ -914,6 +1011,7 @@ class Up2k(object):
|
|||||||
return self._orz(db_path)
|
return self._orz(db_path)
|
||||||
|
|
||||||
def handle_json(self, cj):
|
def handle_json(self, cj):
|
||||||
|
with self.mutex:
|
||||||
if not self.register_vpath(cj["ptop"], cj["vcfg"]):
|
if not self.register_vpath(cj["ptop"], cj["vcfg"]):
|
||||||
if cj["ptop"] not in self.registry:
|
if cj["ptop"] not in self.registry:
|
||||||
raise Pebkac(410, "location unavailable")
|
raise Pebkac(410, "location unavailable")
|
||||||
@@ -939,7 +1037,7 @@ class Up2k(object):
|
|||||||
if dp_dir.startswith("//") or dp_fn.startswith("//"):
|
if dp_dir.startswith("//") or dp_fn.startswith("//"):
|
||||||
dp_dir, dp_fn = s3dec(dp_dir, dp_fn)
|
dp_dir, dp_fn = s3dec(dp_dir, dp_fn)
|
||||||
|
|
||||||
dp_abs = os.path.join(cj["ptop"], dp_dir, dp_fn).replace("\\", "/")
|
dp_abs = "/".join([cj["ptop"], dp_dir, dp_fn])
|
||||||
# relying on path.exists to return false on broken symlinks
|
# relying on path.exists to return false on broken symlinks
|
||||||
if os.path.exists(fsenc(dp_abs)):
|
if os.path.exists(fsenc(dp_abs)):
|
||||||
job = {
|
job = {
|
||||||
@@ -965,7 +1063,7 @@ class Up2k(object):
|
|||||||
for fn in names:
|
for fn in names:
|
||||||
path = os.path.join(job["ptop"], job["prel"], fn)
|
path = os.path.join(job["ptop"], job["prel"], fn)
|
||||||
try:
|
try:
|
||||||
if os.path.getsize(path) > 0:
|
if os.path.getsize(fsenc(path)) > 0:
|
||||||
# upload completed or both present
|
# upload completed or both present
|
||||||
break
|
break
|
||||||
except:
|
except:
|
||||||
@@ -1089,6 +1187,9 @@ class Up2k(object):
|
|||||||
raise Pebkac(400, "unknown wark")
|
raise Pebkac(400, "unknown wark")
|
||||||
|
|
||||||
if chash not in job["need"]:
|
if chash not in job["need"]:
|
||||||
|
msg = "chash = {} , need:\n".format(chash)
|
||||||
|
msg += "\n".join(job["need"])
|
||||||
|
self.log(msg)
|
||||||
raise Pebkac(400, "already got that but thanks??")
|
raise Pebkac(400, "already got that but thanks??")
|
||||||
|
|
||||||
nchunk = [n for n, v in enumerate(job["hash"]) if v == chash]
|
nchunk = [n for n, v in enumerate(job["hash"]) if v == chash]
|
||||||
@@ -1154,6 +1255,7 @@ class Up2k(object):
|
|||||||
|
|
||||||
if "e2t" in self.flags[ptop]:
|
if "e2t" in self.flags[ptop]:
|
||||||
self.tagq.put([ptop, wark, rd, fn])
|
self.tagq.put([ptop, wark, rd, fn])
|
||||||
|
self.n_tagq += 1
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
@@ -1194,12 +1296,15 @@ class Up2k(object):
|
|||||||
return wark
|
return wark
|
||||||
|
|
||||||
def _hashlist_from_file(self, path):
|
def _hashlist_from_file(self, path):
|
||||||
fsz = os.path.getsize(path)
|
pp = self.pp if hasattr(self, "pp") else None
|
||||||
|
fsz = os.path.getsize(fsenc(path))
|
||||||
csz = up2k_chunksize(fsz)
|
csz = up2k_chunksize(fsz)
|
||||||
ret = []
|
ret = []
|
||||||
with open(path, "rb", 512 * 1024) as f:
|
with open(fsenc(path), "rb", 512 * 1024) as f:
|
||||||
while fsz > 0:
|
while fsz > 0:
|
||||||
self.pp.msg = "{} MB, {}".format(int(fsz / 1024 / 1024), path)
|
if pp:
|
||||||
|
pp.msg = "{} MB, {}".format(int(fsz / 1024 / 1024), path)
|
||||||
|
|
||||||
hashobj = hashlib.sha512()
|
hashobj = hashlib.sha512()
|
||||||
rem = min(csz, fsz)
|
rem = min(csz, fsz)
|
||||||
fsz -= rem
|
fsz -= rem
|
||||||
@@ -1276,11 +1381,12 @@ class Up2k(object):
|
|||||||
for k, reg in self.registry.items():
|
for k, reg in self.registry.items():
|
||||||
self._snap_reg(prev, k, reg, discard_interval)
|
self._snap_reg(prev, k, reg, discard_interval)
|
||||||
|
|
||||||
def _snap_reg(self, prev, k, reg, discard_interval):
|
def _snap_reg(self, prev, ptop, reg, discard_interval):
|
||||||
now = time.time()
|
now = time.time()
|
||||||
|
histpath = self.asrv.vfs.histtab[ptop]
|
||||||
rm = [x for x in reg.values() if now - x["poke"] > discard_interval]
|
rm = [x for x in reg.values() if now - x["poke"] > discard_interval]
|
||||||
if rm:
|
if rm:
|
||||||
m = "dropping {} abandoned uploads in {}".format(len(rm), k)
|
m = "dropping {} abandoned uploads in {}".format(len(rm), ptop)
|
||||||
vis = [self._vis_job_progress(x) for x in rm]
|
vis = [self._vis_job_progress(x) for x in rm]
|
||||||
self.log("\n".join([m] + vis))
|
self.log("\n".join([m] + vis))
|
||||||
for job in rm:
|
for job in rm:
|
||||||
@@ -1288,31 +1394,31 @@ class Up2k(object):
|
|||||||
try:
|
try:
|
||||||
# remove the filename reservation
|
# remove the filename reservation
|
||||||
path = os.path.join(job["ptop"], job["prel"], job["name"])
|
path = os.path.join(job["ptop"], job["prel"], job["name"])
|
||||||
if os.path.getsize(path) == 0:
|
if os.path.getsize(fsenc(path)) == 0:
|
||||||
os.unlink(path)
|
os.unlink(fsenc(path))
|
||||||
|
|
||||||
if len(job["hash"]) == len(job["need"]):
|
if len(job["hash"]) == len(job["need"]):
|
||||||
# PARTIAL is empty, delete that too
|
# PARTIAL is empty, delete that too
|
||||||
path = os.path.join(job["ptop"], job["prel"], job["tnam"])
|
path = os.path.join(job["ptop"], job["prel"], job["tnam"])
|
||||||
os.unlink(path)
|
os.unlink(fsenc(path))
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
path = os.path.join(k, ".hist", "up2k.snap")
|
path = os.path.join(histpath, "up2k.snap")
|
||||||
if not reg:
|
if not reg:
|
||||||
if k not in prev or prev[k] is not None:
|
if ptop not in prev or prev[ptop] is not None:
|
||||||
prev[k] = None
|
prev[ptop] = None
|
||||||
if os.path.exists(path):
|
if os.path.exists(fsenc(path)):
|
||||||
os.unlink(path)
|
os.unlink(fsenc(path))
|
||||||
return
|
return
|
||||||
|
|
||||||
newest = max(x["poke"] for _, x in reg.items()) if reg else 0
|
newest = max(x["poke"] for _, x in reg.items()) if reg else 0
|
||||||
etag = [len(reg), newest]
|
etag = [len(reg), newest]
|
||||||
if etag == prev.get(k, None):
|
if etag == prev.get(ptop, None):
|
||||||
return
|
return
|
||||||
|
|
||||||
try:
|
try:
|
||||||
os.mkdir(os.path.join(k, ".hist"))
|
os.makedirs(histpath)
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@@ -1324,18 +1430,25 @@ class Up2k(object):
|
|||||||
atomic_move(path2, path)
|
atomic_move(path2, path)
|
||||||
|
|
||||||
self.log("snap: {} |{}|".format(path, len(reg.keys())))
|
self.log("snap: {} |{}|".format(path, len(reg.keys())))
|
||||||
prev[k] = etag
|
prev[ptop] = etag
|
||||||
|
|
||||||
def _tagger(self):
|
def _tagger(self):
|
||||||
|
with self.mutex:
|
||||||
|
self.n_tagq += 1
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
|
with self.mutex:
|
||||||
|
self.n_tagq -= 1
|
||||||
|
|
||||||
ptop, wark, rd, fn = self.tagq.get()
|
ptop, wark, rd, fn = self.tagq.get()
|
||||||
if "e2t" not in self.flags[ptop]:
|
if "e2t" not in self.flags[ptop]:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
# self.log("\n " + repr([ptop, rd, fn]))
|
||||||
abspath = os.path.join(ptop, rd, fn)
|
abspath = os.path.join(ptop, rd, fn)
|
||||||
tags = self.mtag.get(abspath)
|
tags = self.mtag.get(abspath)
|
||||||
ntags1 = len(tags)
|
ntags1 = len(tags)
|
||||||
parsers = self._get_parsers(ptop, tags)
|
parsers = self._get_parsers(ptop, tags, abspath)
|
||||||
if parsers:
|
if parsers:
|
||||||
tags.update(self.mtag.get_bin(parsers, abspath))
|
tags.update(self.mtag.get_bin(parsers, abspath))
|
||||||
|
|
||||||
@@ -1357,8 +1470,16 @@ class Up2k(object):
|
|||||||
self.log("tagged {} ({}+{})".format(abspath, ntags1, len(tags) - ntags1))
|
self.log("tagged {} ({}+{})".format(abspath, ntags1, len(tags) - ntags1))
|
||||||
|
|
||||||
def _hasher(self):
|
def _hasher(self):
|
||||||
|
with self.mutex:
|
||||||
|
self.n_hashq += 1
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
|
with self.mutex:
|
||||||
|
self.n_hashq -= 1
|
||||||
|
# self.log("hashq {}".format(self.n_hashq))
|
||||||
|
|
||||||
ptop, rd, fn = self.hashq.get()
|
ptop, rd, fn = self.hashq.get()
|
||||||
|
# self.log("hashq {} pop {}/{}/{}".format(self.n_hashq, ptop, rd, fn))
|
||||||
if "e2d" not in self.flags[ptop]:
|
if "e2d" not in self.flags[ptop]:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
@@ -1371,8 +1492,11 @@ class Up2k(object):
|
|||||||
self.idx_wark(ptop, wark, rd, fn, inf.st_mtime, inf.st_size)
|
self.idx_wark(ptop, wark, rd, fn, inf.st_mtime, inf.st_size)
|
||||||
|
|
||||||
def hash_file(self, ptop, flags, rd, fn):
|
def hash_file(self, ptop, flags, rd, fn):
|
||||||
|
with self.mutex:
|
||||||
self.register_vpath(ptop, flags)
|
self.register_vpath(ptop, flags)
|
||||||
self.hashq.put([ptop, rd, fn])
|
self.hashq.put([ptop, rd, fn])
|
||||||
|
self.n_hashq += 1
|
||||||
|
# self.log("hashq {} push {}/{}/{}".format(self.n_hashq, ptop, rd, fn))
|
||||||
|
|
||||||
|
|
||||||
def up2k_chunksize(filesize):
|
def up2k_chunksize(filesize):
|
||||||
@@ -1394,9 +1518,12 @@ def up2k_wark_from_hashlist(salt, filesize, hashes):
|
|||||||
ident.extend(hashes)
|
ident.extend(hashes)
|
||||||
ident = "\n".join(ident)
|
ident = "\n".join(ident)
|
||||||
|
|
||||||
hasher = hashlib.sha512()
|
wark = hashlib.sha512(ident.encode("utf-8")).digest()
|
||||||
hasher.update(ident.encode("utf-8"))
|
wark = base64.urlsafe_b64encode(wark)
|
||||||
digest = hasher.digest()[:32]
|
return wark.decode("ascii")[:43]
|
||||||
|
|
||||||
wark = base64.urlsafe_b64encode(digest)
|
|
||||||
return wark.decode("utf-8").rstrip("=")
|
def up2k_wark_from_metadata(salt, sz, lastmod, rd, fn):
|
||||||
|
ret = fsenc("{}\n{}\n{}\n{}\n{}".format(salt, lastmod, sz, rd, fn))
|
||||||
|
ret = base64.urlsafe_b64encode(hashlib.sha512(ret).digest())
|
||||||
|
return "#{}".format(ret[:42].decode("ascii"))
|
||||||
|
|||||||
@@ -15,6 +15,7 @@ import threading
|
|||||||
import mimetypes
|
import mimetypes
|
||||||
import contextlib
|
import contextlib
|
||||||
import subprocess as sp # nosec
|
import subprocess as sp # nosec
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
from .__init__ import PY2, WINDOWS, ANYWIN
|
from .__init__ import PY2, WINDOWS, ANYWIN
|
||||||
from .stolen import surrogateescape
|
from .stolen import surrogateescape
|
||||||
@@ -34,10 +35,12 @@ if not PY2:
|
|||||||
from urllib.parse import unquote_to_bytes as unquote
|
from urllib.parse import unquote_to_bytes as unquote
|
||||||
from urllib.parse import quote_from_bytes as quote
|
from urllib.parse import quote_from_bytes as quote
|
||||||
from queue import Queue
|
from queue import Queue
|
||||||
|
from io import BytesIO
|
||||||
else:
|
else:
|
||||||
from urllib import unquote # pylint: disable=no-name-in-module
|
from urllib import unquote # pylint: disable=no-name-in-module
|
||||||
from urllib import quote # pylint: disable=no-name-in-module
|
from urllib import quote # pylint: disable=no-name-in-module
|
||||||
from Queue import Queue # pylint: disable=import-error,no-name-in-module
|
from Queue import Queue # pylint: disable=import-error,no-name-in-module
|
||||||
|
from StringIO import StringIO as BytesIO
|
||||||
|
|
||||||
surrogateescape.register_surrogateescape()
|
surrogateescape.register_surrogateescape()
|
||||||
FS_ENCODING = sys.getfilesystemencoding()
|
FS_ENCODING = sys.getfilesystemencoding()
|
||||||
@@ -45,6 +48,9 @@ if WINDOWS and PY2:
|
|||||||
FS_ENCODING = "utf-8"
|
FS_ENCODING = "utf-8"
|
||||||
|
|
||||||
|
|
||||||
|
HTTP_TS_FMT = "%a, %d %b %Y %H:%M:%S GMT"
|
||||||
|
|
||||||
|
|
||||||
HTTPCODE = {
|
HTTPCODE = {
|
||||||
200: "OK",
|
200: "OK",
|
||||||
204: "No Content",
|
204: "No Content",
|
||||||
@@ -73,6 +79,13 @@ IMPLICATIONS = [
|
|||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
MIMES = {
|
||||||
|
"md": "text/plain; charset=UTF-8",
|
||||||
|
"opus": "audio/ogg; codecs=opus",
|
||||||
|
"webp": "image/webp",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
REKOBO_KEY = {
|
REKOBO_KEY = {
|
||||||
v: ln.split(" ", 1)[0]
|
v: ln.split(" ", 1)[0]
|
||||||
for ln in """
|
for ln in """
|
||||||
@@ -124,6 +137,32 @@ class Counter(object):
|
|||||||
self.v = absval
|
self.v = absval
|
||||||
|
|
||||||
|
|
||||||
|
class Cooldown(object):
|
||||||
|
def __init__(self, maxage):
|
||||||
|
self.maxage = maxage
|
||||||
|
self.mutex = threading.Lock()
|
||||||
|
self.hist = {}
|
||||||
|
self.oldest = 0
|
||||||
|
|
||||||
|
def poke(self, key):
|
||||||
|
with self.mutex:
|
||||||
|
now = time.time()
|
||||||
|
|
||||||
|
ret = False
|
||||||
|
v = self.hist.get(key, 0)
|
||||||
|
if now - v > self.maxage:
|
||||||
|
self.hist[key] = now
|
||||||
|
ret = True
|
||||||
|
|
||||||
|
if self.oldest - now > self.maxage * 2:
|
||||||
|
self.hist = {
|
||||||
|
k: v for k, v in self.hist.items() if now - v < self.maxage
|
||||||
|
}
|
||||||
|
self.oldest = sorted(self.hist.values())[0]
|
||||||
|
|
||||||
|
return ret
|
||||||
|
|
||||||
|
|
||||||
class Unrecv(object):
|
class Unrecv(object):
|
||||||
"""
|
"""
|
||||||
undo any number of socket recv ops
|
undo any number of socket recv ops
|
||||||
@@ -154,7 +193,7 @@ class ProgressPrinter(threading.Thread):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
threading.Thread.__init__(self)
|
threading.Thread.__init__(self, name="pp")
|
||||||
self.daemon = True
|
self.daemon = True
|
||||||
self.msg = None
|
self.msg = None
|
||||||
self.end = False
|
self.end = False
|
||||||
@@ -169,6 +208,8 @@ class ProgressPrinter(threading.Thread):
|
|||||||
|
|
||||||
msg = self.msg
|
msg = self.msg
|
||||||
uprint(" {}\033[K\r".format(msg))
|
uprint(" {}\033[K\r".format(msg))
|
||||||
|
if PY2:
|
||||||
|
sys.stdout.flush()
|
||||||
|
|
||||||
print("\033[K", end="")
|
print("\033[K", end="")
|
||||||
sys.stdout.flush() # necessary on win10 even w/ stderr btw
|
sys.stdout.flush() # necessary on win10 even w/ stderr btw
|
||||||
@@ -213,6 +254,17 @@ def trace(*args, **kwargs):
|
|||||||
nuprint(msg)
|
nuprint(msg)
|
||||||
|
|
||||||
|
|
||||||
|
def min_ex():
|
||||||
|
et, ev, tb = sys.exc_info()
|
||||||
|
tb = traceback.extract_tb(tb, 2)
|
||||||
|
ex = [
|
||||||
|
"{} @ {} <{}>: {}".format(fp.split(os.sep)[-1], ln, fun, txt)
|
||||||
|
for fp, ln, fun, txt in tb
|
||||||
|
]
|
||||||
|
ex.append("{}: {}".format(et.__name__, ev))
|
||||||
|
return "\n".join(ex)
|
||||||
|
|
||||||
|
|
||||||
@contextlib.contextmanager
|
@contextlib.contextmanager
|
||||||
def ren_open(fname, *args, **kwargs):
|
def ren_open(fname, *args, **kwargs):
|
||||||
fdir = kwargs.pop("fdir", None)
|
fdir = kwargs.pop("fdir", None)
|
||||||
@@ -223,6 +275,11 @@ def ren_open(fname, *args, **kwargs):
|
|||||||
yield {"orz": [f, fname]}
|
yield {"orz": [f, fname]}
|
||||||
return
|
return
|
||||||
|
|
||||||
|
if suffix:
|
||||||
|
ext = fname.split(".")[-1]
|
||||||
|
if len(ext) < 7:
|
||||||
|
suffix += "." + ext
|
||||||
|
|
||||||
orig_name = fname
|
orig_name = fname
|
||||||
bname = fname
|
bname = fname
|
||||||
ext = ""
|
ext = ""
|
||||||
@@ -243,7 +300,7 @@ def ren_open(fname, *args, **kwargs):
|
|||||||
else:
|
else:
|
||||||
fpath = fname
|
fpath = fname
|
||||||
|
|
||||||
if suffix and os.path.exists(fpath):
|
if suffix and os.path.exists(fsenc(fpath)):
|
||||||
fpath += suffix
|
fpath += suffix
|
||||||
fname += suffix
|
fname += suffix
|
||||||
ext += suffix
|
ext += suffix
|
||||||
@@ -522,8 +579,10 @@ def read_header(sr):
|
|||||||
else:
|
else:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
if len(ret) > ofs + 4:
|
||||||
sr.unrecv(ret[ofs + 4 :])
|
sr.unrecv(ret[ofs + 4 :])
|
||||||
return ret[:ofs].decode("utf-8", "surrogateescape").split("\r\n")
|
|
||||||
|
return ret[:ofs].decode("utf-8", "surrogateescape").lstrip("\r\n").split("\r\n")
|
||||||
|
|
||||||
|
|
||||||
def humansize(sz, terse=False):
|
def humansize(sz, terse=False):
|
||||||
@@ -617,6 +676,11 @@ def exclude_dotfiles(filepaths):
|
|||||||
return [x for x in filepaths if not x.split("/")[-1].startswith(".")]
|
return [x for x in filepaths if not x.split("/")[-1].startswith(".")]
|
||||||
|
|
||||||
|
|
||||||
|
def http_ts(ts):
|
||||||
|
file_dt = datetime.utcfromtimestamp(ts)
|
||||||
|
return file_dt.strftime(HTTP_TS_FMT)
|
||||||
|
|
||||||
|
|
||||||
def html_escape(s, quote=False, crlf=False):
|
def html_escape(s, quote=False, crlf=False):
|
||||||
"""html.escape but also newlines"""
|
"""html.escape but also newlines"""
|
||||||
s = s.replace("&", "&").replace("<", "<").replace(">", ">")
|
s = s.replace("&", "&").replace("<", "<").replace(">", ">")
|
||||||
@@ -723,6 +787,8 @@ def s3dec(rd, fn):
|
|||||||
|
|
||||||
|
|
||||||
def atomic_move(src, dst):
|
def atomic_move(src, dst):
|
||||||
|
src = fsenc(src)
|
||||||
|
dst = fsenc(dst)
|
||||||
if not PY2:
|
if not PY2:
|
||||||
os.replace(src, dst)
|
os.replace(src, dst)
|
||||||
else:
|
else:
|
||||||
@@ -801,13 +867,14 @@ def yieldfile(fn):
|
|||||||
|
|
||||||
|
|
||||||
def hashcopy(actor, fin, fout):
|
def hashcopy(actor, fin, fout):
|
||||||
u32_lim = int((2 ** 31) * 0.9)
|
is_mp = actor.is_mp
|
||||||
hashobj = hashlib.sha512()
|
hashobj = hashlib.sha512()
|
||||||
tlen = 0
|
tlen = 0
|
||||||
for buf in fin:
|
for buf in fin:
|
||||||
|
if is_mp:
|
||||||
actor.workload += 1
|
actor.workload += 1
|
||||||
if actor.workload > u32_lim:
|
if actor.workload > 2 ** 31:
|
||||||
actor.workload = 100 # prevent overflow
|
actor.workload = 100
|
||||||
|
|
||||||
tlen += len(buf)
|
tlen += len(buf)
|
||||||
hashobj.update(buf)
|
hashobj.update(buf)
|
||||||
@@ -819,12 +886,17 @@ def hashcopy(actor, fin, fout):
|
|||||||
return tlen, hashobj.hexdigest(), digest_b64
|
return tlen, hashobj.hexdigest(), digest_b64
|
||||||
|
|
||||||
|
|
||||||
def sendfile_py(lower, upper, f, s):
|
def sendfile_py(lower, upper, f, s, actor=None):
|
||||||
remains = upper - lower
|
remains = upper - lower
|
||||||
f.seek(lower)
|
f.seek(lower)
|
||||||
while remains > 0:
|
while remains > 0:
|
||||||
|
if actor:
|
||||||
|
actor.workload += 1
|
||||||
|
if actor.workload > 2 ** 31:
|
||||||
|
actor.workload = 100
|
||||||
|
|
||||||
# time.sleep(0.01)
|
# time.sleep(0.01)
|
||||||
buf = f.read(min(4096, remains))
|
buf = f.read(min(1024 * 32, remains))
|
||||||
if not buf:
|
if not buf:
|
||||||
return remains
|
return remains
|
||||||
|
|
||||||
@@ -914,18 +986,20 @@ def unescape_cookie(orig):
|
|||||||
return ret
|
return ret
|
||||||
|
|
||||||
|
|
||||||
def guess_mime(url):
|
def guess_mime(url, fallback="application/octet-stream"):
|
||||||
if url.endswith(".md"):
|
try:
|
||||||
return ["text/plain; charset=UTF-8"]
|
_, ext = url.rsplit(".", 1)
|
||||||
|
except:
|
||||||
|
return fallback
|
||||||
|
|
||||||
return mimetypes.guess_type(url)
|
return MIMES.get(ext) or mimetypes.guess_type(url)[0] or fallback
|
||||||
|
|
||||||
|
|
||||||
def runcmd(*argv):
|
def runcmd(*argv):
|
||||||
p = sp.Popen(argv, stdout=sp.PIPE, stderr=sp.PIPE)
|
p = sp.Popen(argv, stdout=sp.PIPE, stderr=sp.PIPE)
|
||||||
stdout, stderr = p.communicate()
|
stdout, stderr = p.communicate()
|
||||||
stdout = stdout.decode("utf-8")
|
stdout = stdout.decode("utf-8", "replace")
|
||||||
stderr = stderr.decode("utf-8")
|
stderr = stderr.decode("utf-8", "replace")
|
||||||
return [p.returncode, stdout, stderr]
|
return [p.returncode, stdout, stderr]
|
||||||
|
|
||||||
|
|
||||||
@@ -937,6 +1011,17 @@ def chkcmd(*argv):
|
|||||||
return sout, serr
|
return sout, serr
|
||||||
|
|
||||||
|
|
||||||
|
def mchkcmd(argv, timeout=10):
|
||||||
|
if PY2:
|
||||||
|
with open(os.devnull, "wb") as f:
|
||||||
|
rv = sp.call(argv, stdout=f, stderr=f)
|
||||||
|
else:
|
||||||
|
rv = sp.call(argv, stdout=sp.DEVNULL, stderr=sp.DEVNULL, timeout=timeout)
|
||||||
|
|
||||||
|
if rv:
|
||||||
|
raise sp.CalledProcessError(rv, (argv[0], b"...", argv[-1]))
|
||||||
|
|
||||||
|
|
||||||
def gzip_orig_sz(fn):
|
def gzip_orig_sz(fn):
|
||||||
with open(fsenc(fn), "rb") as f:
|
with open(fsenc(fn), "rb") as f:
|
||||||
f.seek(-4, 2)
|
f.seek(-4, 2)
|
||||||
|
|||||||
@@ -1,3 +1,6 @@
|
|||||||
|
:root {
|
||||||
|
--grid-sz: 10em;
|
||||||
|
}
|
||||||
* {
|
* {
|
||||||
line-height: 1.2em;
|
line-height: 1.2em;
|
||||||
}
|
}
|
||||||
@@ -64,6 +67,11 @@ a, #files tbody div a:last-child {
|
|||||||
background: #161616;
|
background: #161616;
|
||||||
text-decoration: underline;
|
text-decoration: underline;
|
||||||
}
|
}
|
||||||
|
#files thead {
|
||||||
|
background: #333;
|
||||||
|
position: sticky;
|
||||||
|
top: 0;
|
||||||
|
}
|
||||||
#files thead a {
|
#files thead a {
|
||||||
color: #999;
|
color: #999;
|
||||||
font-weight: normal;
|
font-weight: normal;
|
||||||
@@ -180,11 +188,32 @@ a, #files tbody div a:last-child {
|
|||||||
color: #840;
|
color: #840;
|
||||||
text-shadow: 0 0 .3em #b80;
|
text-shadow: 0 0 .3em #b80;
|
||||||
}
|
}
|
||||||
#files tbody tr.sel td {
|
#files tbody tr.sel td,
|
||||||
|
#ggrid a.sel,
|
||||||
|
html.light #ggrid a.sel {
|
||||||
color: #fff;
|
color: #fff;
|
||||||
background: #925;
|
background: #925;
|
||||||
border-color: #c37;
|
border-color: #c37;
|
||||||
}
|
}
|
||||||
|
#files tbody tr.sel:hover td,
|
||||||
|
#ggrid a.sel:hover,
|
||||||
|
html.light #ggrid a.sel:hover {
|
||||||
|
color: #fff;
|
||||||
|
background: #d39;
|
||||||
|
border-color: #d48;
|
||||||
|
text-shadow: 1px 1px 0 #804;
|
||||||
|
}
|
||||||
|
#ggrid a.sel,
|
||||||
|
html.light #ggrid a.sel {
|
||||||
|
border-top: 1px solid #d48;
|
||||||
|
box-shadow: 0 .1em 1.2em #b36;
|
||||||
|
transition: all 0.2s cubic-bezier(.2, 2.2, .5, 1); /* https://cubic-bezier.com/#.4,2,.7,1 */
|
||||||
|
}
|
||||||
|
#ggrid a.sel img {
|
||||||
|
opacity: .7;
|
||||||
|
box-shadow: 0 0 1em #b36;
|
||||||
|
filter: contrast(130%) brightness(107%);
|
||||||
|
}
|
||||||
#files tr.sel a {
|
#files tr.sel a {
|
||||||
color: #fff;
|
color: #fff;
|
||||||
}
|
}
|
||||||
@@ -252,7 +281,10 @@ a, #files tbody div a:last-child {
|
|||||||
background: #3c3c3c;
|
background: #3c3c3c;
|
||||||
}
|
}
|
||||||
#wtico {
|
#wtico {
|
||||||
cursor: url(/.cpr/dd/1.png), pointer;
|
cursor: url(/.cpr/dd/4.png), pointer;
|
||||||
|
animation: cursor 500ms;
|
||||||
|
}
|
||||||
|
#wtico:hover {
|
||||||
animation: cursor 500ms infinite;
|
animation: cursor 500ms infinite;
|
||||||
}
|
}
|
||||||
@keyframes cursor {
|
@keyframes cursor {
|
||||||
@@ -260,7 +292,7 @@ a, #files tbody div a:last-child {
|
|||||||
30% {cursor: url(/.cpr/dd/3.png), pointer}
|
30% {cursor: url(/.cpr/dd/3.png), pointer}
|
||||||
50% {cursor: url(/.cpr/dd/4.png), pointer}
|
50% {cursor: url(/.cpr/dd/4.png), pointer}
|
||||||
75% {cursor: url(/.cpr/dd/5.png), pointer}
|
75% {cursor: url(/.cpr/dd/5.png), pointer}
|
||||||
85% {cursor: url(/.cpr/dd/1.png), pointer}
|
85% {cursor: url(/.cpr/dd/4.png), pointer}
|
||||||
}
|
}
|
||||||
@keyframes spin {
|
@keyframes spin {
|
||||||
100% {transform: rotate(360deg)}
|
100% {transform: rotate(360deg)}
|
||||||
@@ -281,29 +313,48 @@ a, #files tbody div a:last-child {
|
|||||||
padding: .2em 0 0 .07em;
|
padding: .2em 0 0 .07em;
|
||||||
color: #fff;
|
color: #fff;
|
||||||
}
|
}
|
||||||
#wzip {
|
#wzip, #wnp {
|
||||||
display: none;
|
display: none;
|
||||||
margin-right: .3em;
|
margin-right: .3em;
|
||||||
padding-right: .3em;
|
padding-right: .3em;
|
||||||
border-right: .1em solid #555;
|
border-right: .1em solid #555;
|
||||||
}
|
}
|
||||||
|
#wnp a {
|
||||||
|
position: relative;
|
||||||
|
font-size: .47em;
|
||||||
|
margin: 0 .1em;
|
||||||
|
top: -.4em;
|
||||||
|
}
|
||||||
|
#wnp a+a {
|
||||||
|
margin-left: .33em;
|
||||||
|
}
|
||||||
#wtoggle,
|
#wtoggle,
|
||||||
#wtoggle * {
|
#wtoggle * {
|
||||||
line-height: 1em;
|
line-height: 1em;
|
||||||
}
|
}
|
||||||
|
#wtoggle.np {
|
||||||
|
width: 5.5em;
|
||||||
|
}
|
||||||
#wtoggle.sel {
|
#wtoggle.sel {
|
||||||
width: 6.4em;
|
width: 6.4em;
|
||||||
}
|
}
|
||||||
#wtoggle.sel #wzip {
|
#wtoggle.sel #wzip,
|
||||||
|
#wtoggle.np #wnp {
|
||||||
display: inline-block;
|
display: inline-block;
|
||||||
}
|
}
|
||||||
#wtoggle.sel #wzip a {
|
#wtoggle.sel.np #wnp {
|
||||||
|
display: none;
|
||||||
|
}
|
||||||
|
#wzip a {
|
||||||
font-size: .4em;
|
font-size: .4em;
|
||||||
padding: 0 .3em;
|
padding: 0 .3em;
|
||||||
margin: -.3em .2em;
|
margin: -.3em .2em;
|
||||||
position: relative;
|
position: relative;
|
||||||
display: inline-block;
|
display: inline-block;
|
||||||
}
|
}
|
||||||
|
#wzip a+a {
|
||||||
|
margin-left: .8em;
|
||||||
|
}
|
||||||
#wtoggle.sel #wzip #selzip {
|
#wtoggle.sel #wzip #selzip {
|
||||||
top: -.6em;
|
top: -.6em;
|
||||||
padding: .4em .3em;
|
padding: .4em .3em;
|
||||||
@@ -478,6 +529,17 @@ input[type="checkbox"]:checked+label {
|
|||||||
height: 1em;
|
height: 1em;
|
||||||
margin: .2em 0 -1em 1.6em;
|
margin: .2em 0 -1em 1.6em;
|
||||||
}
|
}
|
||||||
|
#tq_raw {
|
||||||
|
width: calc(100% - 2em);
|
||||||
|
margin: .3em 0 0 1.4em;
|
||||||
|
}
|
||||||
|
#tq_raw td+td {
|
||||||
|
width: 100%;
|
||||||
|
}
|
||||||
|
#op_search #q_raw {
|
||||||
|
width: 100%;
|
||||||
|
display: block;
|
||||||
|
}
|
||||||
#files td div span {
|
#files td div span {
|
||||||
color: #fff;
|
color: #fff;
|
||||||
padding: 0 .4em;
|
padding: 0 .4em;
|
||||||
@@ -508,7 +570,6 @@ input[type="checkbox"]:checked+label {
|
|||||||
left: 0;
|
left: 0;
|
||||||
bottom: 0;
|
bottom: 0;
|
||||||
top: 7em;
|
top: 7em;
|
||||||
padding-top: .2em;
|
|
||||||
overflow-y: auto;
|
overflow-y: auto;
|
||||||
-ms-scroll-chaining: none;
|
-ms-scroll-chaining: none;
|
||||||
overscroll-behavior-y: none;
|
overscroll-behavior-y: none;
|
||||||
@@ -532,8 +593,7 @@ input[type="checkbox"]:checked+label {
|
|||||||
left: -1.7em;
|
left: -1.7em;
|
||||||
width: calc(100% + 1.3em);
|
width: calc(100% + 1.3em);
|
||||||
}
|
}
|
||||||
.tglbtn,
|
.btn {
|
||||||
#tree>a+a {
|
|
||||||
padding: .2em .4em;
|
padding: .2em .4em;
|
||||||
font-size: 1.2em;
|
font-size: 1.2em;
|
||||||
background: #2a2a2a;
|
background: #2a2a2a;
|
||||||
@@ -543,12 +603,10 @@ input[type="checkbox"]:checked+label {
|
|||||||
position: relative;
|
position: relative;
|
||||||
top: -.2em;
|
top: -.2em;
|
||||||
}
|
}
|
||||||
.tglbtn:hover,
|
.btn:hover {
|
||||||
#tree>a+a:hover {
|
|
||||||
background: #805;
|
background: #805;
|
||||||
}
|
}
|
||||||
.tglbtn.on,
|
.tgl.btn.on {
|
||||||
#tree>a+a.on {
|
|
||||||
background: #fc4;
|
background: #fc4;
|
||||||
color: #400;
|
color: #400;
|
||||||
text-shadow: none;
|
text-shadow: none;
|
||||||
@@ -556,6 +614,7 @@ input[type="checkbox"]:checked+label {
|
|||||||
#detree {
|
#detree {
|
||||||
padding: .3em .5em;
|
padding: .3em .5em;
|
||||||
font-size: 1.5em;
|
font-size: 1.5em;
|
||||||
|
line-height: 1.5em;
|
||||||
}
|
}
|
||||||
#tree ul,
|
#tree ul,
|
||||||
#tree li {
|
#tree li {
|
||||||
@@ -585,7 +644,6 @@ input[type="checkbox"]:checked+label {
|
|||||||
}
|
}
|
||||||
#treeul a+a {
|
#treeul a+a {
|
||||||
width: calc(100% - 2em);
|
width: calc(100% - 2em);
|
||||||
background: #333;
|
|
||||||
line-height: 1em;
|
line-height: 1em;
|
||||||
}
|
}
|
||||||
#treeul a+a:hover {
|
#treeul a+a:hover {
|
||||||
@@ -692,6 +750,76 @@ input[type="checkbox"]:checked+label {
|
|||||||
font-family: monospace, monospace;
|
font-family: monospace, monospace;
|
||||||
line-height: 2em;
|
line-height: 2em;
|
||||||
}
|
}
|
||||||
|
#griden.on+#thumbs {
|
||||||
|
opacity: .3;
|
||||||
|
}
|
||||||
|
#ghead {
|
||||||
|
background: #3c3c3c;
|
||||||
|
border: 1px solid #444;
|
||||||
|
border-radius: .3em;
|
||||||
|
padding: .5em;
|
||||||
|
margin: 0 1.5em 1em .4em;
|
||||||
|
position: sticky;
|
||||||
|
top: -.3em;
|
||||||
|
}
|
||||||
|
html.light #ghead {
|
||||||
|
background: #f7f7f7;
|
||||||
|
border-color: #ddd;
|
||||||
|
}
|
||||||
|
#ghead .btn {
|
||||||
|
position: relative;
|
||||||
|
top: 0;
|
||||||
|
}
|
||||||
|
#ggrid {
|
||||||
|
padding-top: .5em;
|
||||||
|
}
|
||||||
|
#ggrid a {
|
||||||
|
display: inline-block;
|
||||||
|
width: var(--grid-sz);
|
||||||
|
vertical-align: top;
|
||||||
|
overflow-wrap: break-word;
|
||||||
|
background: #383838;
|
||||||
|
border: 1px solid #444;
|
||||||
|
border-top: 1px solid #555;
|
||||||
|
box-shadow: 0 .1em .2em #222;
|
||||||
|
border-radius: .3em;
|
||||||
|
padding: .3em;
|
||||||
|
margin: .5em;
|
||||||
|
}
|
||||||
|
#ggrid a img {
|
||||||
|
border-radius: .2em;
|
||||||
|
max-width: var(--grid-sz);
|
||||||
|
max-height: calc(var(--grid-sz)/1.25);
|
||||||
|
margin: 0 auto;
|
||||||
|
display: block;
|
||||||
|
}
|
||||||
|
#ggrid a span {
|
||||||
|
padding: .2em .3em;
|
||||||
|
display: block;
|
||||||
|
}
|
||||||
|
#ggrid span.dir:before {
|
||||||
|
content: '📂';
|
||||||
|
line-height: 0;
|
||||||
|
font-size: 2em;
|
||||||
|
display: inline-block;
|
||||||
|
margin: -.7em 0 -.5em -.3em;
|
||||||
|
}
|
||||||
|
#ggrid a:hover {
|
||||||
|
background: #444;
|
||||||
|
border-color: #555;
|
||||||
|
color: #fd9;
|
||||||
|
}
|
||||||
|
html.light #ggrid a {
|
||||||
|
background: #f7f7f7;
|
||||||
|
border-color: #ddd;
|
||||||
|
box-shadow: 0 .1em .2em #ddd;
|
||||||
|
}
|
||||||
|
html.light #ggrid a:hover {
|
||||||
|
background: #fff;
|
||||||
|
border-color: #ccc;
|
||||||
|
color: #015;
|
||||||
|
box-shadow: 0 .1em .5em #aaa;
|
||||||
|
}
|
||||||
#pvol,
|
#pvol,
|
||||||
#barbuf,
|
#barbuf,
|
||||||
#barpos,
|
#barpos,
|
||||||
@@ -706,6 +834,21 @@ input[type="checkbox"]:checked+label {
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
html.light {
|
html.light {
|
||||||
color: #333;
|
color: #333;
|
||||||
background: #eee;
|
background: #eee;
|
||||||
@@ -727,18 +870,15 @@ html.light #ops a.act {
|
|||||||
html.light #op_cfg h3 {
|
html.light #op_cfg h3 {
|
||||||
border-color: #ccc;
|
border-color: #ccc;
|
||||||
}
|
}
|
||||||
html.light .tglbtn,
|
html.light .btn {
|
||||||
html.light #tree > a + a {
|
|
||||||
color: #666;
|
color: #666;
|
||||||
background: #ddd;
|
background: #ddd;
|
||||||
box-shadow: none;
|
box-shadow: none;
|
||||||
}
|
}
|
||||||
html.light .tglbtn:hover,
|
html.light .btn:hover {
|
||||||
html.light #tree > a + a:hover {
|
|
||||||
background: #caf;
|
background: #caf;
|
||||||
}
|
}
|
||||||
html.light .tglbtn.on,
|
html.light .tgl.btn.on {
|
||||||
html.light #tree > a + a.on {
|
|
||||||
background: #4a0;
|
background: #4a0;
|
||||||
color: #fff;
|
color: #fff;
|
||||||
}
|
}
|
||||||
@@ -776,6 +916,7 @@ html.light #files {
|
|||||||
}
|
}
|
||||||
html.light #files thead th {
|
html.light #files thead th {
|
||||||
background: #eee;
|
background: #eee;
|
||||||
|
border-radius: 0;
|
||||||
}
|
}
|
||||||
html.light #files tr td {
|
html.light #files tr td {
|
||||||
border-top: 1px solid #ddd;
|
border-top: 1px solid #ddd;
|
||||||
|
|||||||
@@ -8,6 +8,9 @@
|
|||||||
<meta name="viewport" content="width=device-width, initial-scale=0.8">
|
<meta name="viewport" content="width=device-width, initial-scale=0.8">
|
||||||
<link rel="stylesheet" type="text/css" media="screen" href="/.cpr/browser.css{{ ts }}">
|
<link rel="stylesheet" type="text/css" media="screen" href="/.cpr/browser.css{{ ts }}">
|
||||||
<link rel="stylesheet" type="text/css" media="screen" href="/.cpr/upload.css{{ ts }}">
|
<link rel="stylesheet" type="text/css" media="screen" href="/.cpr/upload.css{{ ts }}">
|
||||||
|
{%- if css %}
|
||||||
|
<link rel="stylesheet" type="text/css" media="screen" href="{{ css }}{{ ts }}">
|
||||||
|
{%- endif %}
|
||||||
</head>
|
</head>
|
||||||
|
|
||||||
<body>
|
<body>
|
||||||
@@ -43,6 +46,8 @@
|
|||||||
<div>
|
<div>
|
||||||
<a id="tooltips" class="tgl btn" href="#">tooltips</a>
|
<a id="tooltips" class="tgl btn" href="#">tooltips</a>
|
||||||
<a id="lightmode" class="tgl btn" href="#">lightmode</a>
|
<a id="lightmode" class="tgl btn" href="#">lightmode</a>
|
||||||
|
<a id="griden" class="tgl btn" href="#">the grid</a>
|
||||||
|
<a id="thumbs" class="tgl btn" href="#">thumbs</a>
|
||||||
</div>
|
</div>
|
||||||
{%- if have_zip %}
|
{%- if have_zip %}
|
||||||
<h3>folder download</h3>
|
<h3>folder download</h3>
|
||||||
@@ -61,8 +66,8 @@
|
|||||||
|
|
||||||
<div id="tree">
|
<div id="tree">
|
||||||
<a href="#" id="detree">🍞...</a>
|
<a href="#" id="detree">🍞...</a>
|
||||||
<a href="#" step="2" id="twobytwo">+</a>
|
<a href="#" class="btn" step="2" id="twobytwo">+</a>
|
||||||
<a href="#" step="-2" id="twig">–</a>
|
<a href="#" class="btn" step="-2" id="twig">–</a>
|
||||||
<a href="#" class="tgl btn" id="dyntree">a</a>
|
<a href="#" class="tgl btn" id="dyntree">a</a>
|
||||||
<ul id="treeul"></ul>
|
<ul id="treeul"></ul>
|
||||||
<div id="thx_ff"> </div>
|
<div id="thx_ff"> </div>
|
||||||
@@ -114,22 +119,7 @@
|
|||||||
<div id="srv_info"><span>{{ srv_info }}</span></div>
|
<div id="srv_info"><span>{{ srv_info }}</span></div>
|
||||||
{%- endif %}
|
{%- endif %}
|
||||||
|
|
||||||
<div id="widget">
|
<div id="widget"></div>
|
||||||
<div id="wtoggle">
|
|
||||||
<span id="wzip">
|
|
||||||
<a href="#" id="selall">sel.<br />all</a>
|
|
||||||
<a href="#" id="selinv">sel.<br />inv.</a>
|
|
||||||
<a href="#" id="selzip">zip</a>
|
|
||||||
</span><a
|
|
||||||
href="#" id="wtico">♫</a>
|
|
||||||
</div>
|
|
||||||
<div id="widgeti">
|
|
||||||
<div id="pctl"><a href="#" id="bprev">⏮</a><a href="#" id="bplay">▶</a><a href="#" id="bnext">⏭</a></div>
|
|
||||||
<canvas id="pvol" width="288" height="38"></canvas>
|
|
||||||
<canvas id="barpos"></canvas>
|
|
||||||
<canvas id="barbuf"></canvas>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<script>
|
<script>
|
||||||
var tag_order_cfg = {{ tag_order }};
|
var tag_order_cfg = {{ tag_order }};
|
||||||
|
|||||||
@@ -7,13 +7,47 @@ function dbg(msg) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
// add widget buttons
|
||||||
|
ebi('widget').innerHTML = (
|
||||||
|
'<div id="wtoggle">' +
|
||||||
|
'<span id="wzip"><a' +
|
||||||
|
' href="#" id="selall">sel.<br />all</a><a' +
|
||||||
|
' href="#" id="selinv">sel.<br />inv.</a><a' +
|
||||||
|
' href="#" id="selzip">zip</a>' +
|
||||||
|
'</span><span id="wnp"><a' +
|
||||||
|
' href="#" id="npirc">📋irc</a><a' +
|
||||||
|
' href="#" id="nptxt">📋txt</a>' +
|
||||||
|
'</span><a' +
|
||||||
|
' href="#" id="wtico">♫</a>' +
|
||||||
|
'</div>' +
|
||||||
|
'<div id="widgeti">' +
|
||||||
|
' <div id="pctl"><a href="#" id="bprev">⏮</a><a href="#" id="bplay">▶</a><a href="#" id="bnext">⏭</a></div>' +
|
||||||
|
' <canvas id="pvol" width="288" height="38"></canvas>' +
|
||||||
|
' <canvas id="barpos"></canvas>' +
|
||||||
|
' <canvas id="barbuf"></canvas>' +
|
||||||
|
'</div>'
|
||||||
|
);
|
||||||
|
|
||||||
|
|
||||||
|
var have_webp = null;
|
||||||
|
(function () {
|
||||||
|
var img = new Image();
|
||||||
|
img.onload = function () {
|
||||||
|
have_webp = img.width > 0 && img.height > 0;
|
||||||
|
};
|
||||||
|
img.onerror = function () {
|
||||||
|
have_webp = false;
|
||||||
|
};
|
||||||
|
img.src = "data:image/webp;base64,UklGRhoAAABXRUJQVlA4TA0AAAAvAAAAEAcQERGIiP4HAA==";
|
||||||
|
})();
|
||||||
|
|
||||||
|
|
||||||
// extract songs + add play column
|
// extract songs + add play column
|
||||||
function MPlayer() {
|
function MPlayer() {
|
||||||
this.id = Date.now();
|
this.id = Date.now();
|
||||||
this.au = null;
|
this.au = null;
|
||||||
this.au_native = null;
|
this.au_native = null;
|
||||||
this.au_ogvjs = null;
|
this.au_ogvjs = null;
|
||||||
this.cover_url = '';
|
|
||||||
this.tracks = {};
|
this.tracks = {};
|
||||||
this.order = [];
|
this.order = [];
|
||||||
|
|
||||||
@@ -79,6 +113,8 @@ var widget = (function () {
|
|||||||
var ret = {},
|
var ret = {},
|
||||||
widget = ebi('widget'),
|
widget = ebi('widget'),
|
||||||
wtico = ebi('wtico'),
|
wtico = ebi('wtico'),
|
||||||
|
nptxt = ebi('nptxt'),
|
||||||
|
npirc = ebi('npirc'),
|
||||||
touchmode = false,
|
touchmode = false,
|
||||||
side_open = false,
|
side_open = false,
|
||||||
was_paused = true;
|
was_paused = true;
|
||||||
@@ -116,6 +152,35 @@ var widget = (function () {
|
|||||||
|
|
||||||
return false;
|
return false;
|
||||||
};
|
};
|
||||||
|
npirc.onclick = nptxt.onclick = function (e) {
|
||||||
|
ev(e);
|
||||||
|
var th = ebi('files').tHead.rows[0].cells,
|
||||||
|
tr = QS('#files tr.play').cells,
|
||||||
|
irc = this.getAttribute('id') == 'npirc',
|
||||||
|
ck = irc ? '06' : '',
|
||||||
|
cv = irc ? '07' : '',
|
||||||
|
m = ck + 'np: ';
|
||||||
|
|
||||||
|
for (var a = 1, aa = th.length; a < aa; a++) {
|
||||||
|
var tk = a == 1 ? '' : th[a].getAttribute('name').split('/').slice(-1)[0];
|
||||||
|
var tv = tr[a].getAttribute('html') || tr[a].textContent;
|
||||||
|
m += tk + '(' + cv + tv + ck + ') // ';
|
||||||
|
}
|
||||||
|
|
||||||
|
m += '[' + cv + s2ms(mp.au.currentTime) + ck + '/' + cv + s2ms(mp.au.duration) + ck + ']';
|
||||||
|
|
||||||
|
var o = document.createElement('input');
|
||||||
|
o.style.cssText = 'position:fixed;top:45%;left:48%;padding:1em;z-index:9';
|
||||||
|
o.value = m;
|
||||||
|
document.body.appendChild(o);
|
||||||
|
o.focus();
|
||||||
|
o.select();
|
||||||
|
document.execCommand("copy");
|
||||||
|
o.value = 'copied to clipboard ';
|
||||||
|
setTimeout(function () {
|
||||||
|
document.body.removeChild(o);
|
||||||
|
}, 500);
|
||||||
|
};
|
||||||
return ret;
|
return ret;
|
||||||
})();
|
})();
|
||||||
|
|
||||||
@@ -189,8 +254,8 @@ var pbar = (function () {
|
|||||||
};
|
};
|
||||||
|
|
||||||
r.drawpos = function () {
|
r.drawpos = function () {
|
||||||
if (!mp.au)
|
if (!mp.au || isNaN(mp.au.duration) || isNaN(mp.au.currentTime))
|
||||||
return;
|
return; // not-init || unsupp-codec
|
||||||
|
|
||||||
var bc = r.buf,
|
var bc = r.buf,
|
||||||
pc = r.pos,
|
pc = r.pos,
|
||||||
@@ -513,6 +578,7 @@ function play(tid, seek, call_depth) {
|
|||||||
clmod(trs[a], 'play');
|
clmod(trs[a], 'play');
|
||||||
}
|
}
|
||||||
ebi(oid).parentElement.parentElement.className += ' play';
|
ebi(oid).parentElement.parentElement.className += ' play';
|
||||||
|
clmod(ebi('wtoggle'), 'np', 1);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
if (attempt_play)
|
if (attempt_play)
|
||||||
@@ -642,6 +708,200 @@ function autoplay_blocked(seek) {
|
|||||||
})();
|
})();
|
||||||
|
|
||||||
|
|
||||||
|
var thegrid = (function () {
|
||||||
|
var lfiles = ebi('files');
|
||||||
|
var gfiles = document.createElement('div');
|
||||||
|
gfiles.setAttribute('id', 'gfiles');
|
||||||
|
gfiles.style.display = 'none';
|
||||||
|
gfiles.innerHTML = (
|
||||||
|
'<div id="ghead">' +
|
||||||
|
'<a href="#" class="tgl btn" id="gridsel">multiselect</a> zoom ' +
|
||||||
|
'<a href="#" class="btn" z="-1.2">–</a> ' +
|
||||||
|
'<a href="#" class="btn" z="1.2">+</a> sort by: ' +
|
||||||
|
'<a href="#" s="href">name</a>, ' +
|
||||||
|
'<a href="#" s="sz">size</a>, ' +
|
||||||
|
'<a href="#" s="ts">date</a>, ' +
|
||||||
|
'<a href="#" s="ext">type</a>' +
|
||||||
|
'</div>' +
|
||||||
|
'<div id="ggrid"></div>'
|
||||||
|
);
|
||||||
|
lfiles.parentNode.insertBefore(gfiles, lfiles);
|
||||||
|
|
||||||
|
var r = {
|
||||||
|
'thumbs': bcfg_get('thumbs', true),
|
||||||
|
'en': bcfg_get('griden', false),
|
||||||
|
'sel': bcfg_get('gridsel', false),
|
||||||
|
'sz': fcfg_get('gridsz', 10),
|
||||||
|
'isdirty': true
|
||||||
|
};
|
||||||
|
|
||||||
|
ebi('thumbs').onclick = function (e) {
|
||||||
|
ev(e);
|
||||||
|
r.thumbs = !r.thumbs;
|
||||||
|
bcfg_set('thumbs', r.thumbs);
|
||||||
|
if (r.en) {
|
||||||
|
loadgrid();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
ebi('griden').onclick = function (e) {
|
||||||
|
ev(e);
|
||||||
|
r.en = !r.en;
|
||||||
|
bcfg_set('griden', r.en);
|
||||||
|
if (r.en) {
|
||||||
|
loadgrid();
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
lfiles.style.display = '';
|
||||||
|
gfiles.style.display = 'none';
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
var btnclick = function (e) {
|
||||||
|
ev(e);
|
||||||
|
var s = this.getAttribute('s'),
|
||||||
|
z = this.getAttribute('z');
|
||||||
|
|
||||||
|
if (z)
|
||||||
|
return setsz(z > 0 ? r.sz * z : r.sz / (-z));
|
||||||
|
|
||||||
|
var t = lfiles.tHead.rows[0].cells;
|
||||||
|
for (var a = 0; a < t.length; a++)
|
||||||
|
if (t[a].getAttribute('name') == s) {
|
||||||
|
t[a].click();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
r.setdirty();
|
||||||
|
};
|
||||||
|
|
||||||
|
var links = QSA('#ghead>a');
|
||||||
|
for (var a = 0; a < links.length; a++)
|
||||||
|
links[a].onclick = btnclick;
|
||||||
|
|
||||||
|
ebi('gridsel').onclick = function (e) {
|
||||||
|
ev(e);
|
||||||
|
r.sel = !r.sel;
|
||||||
|
bcfg_set('gridsel', r.sel);
|
||||||
|
r.loadsel();
|
||||||
|
};
|
||||||
|
|
||||||
|
r.setvis = function (vis) {
|
||||||
|
(r.en ? gfiles : lfiles).style.display = vis ? '' : 'none';
|
||||||
|
}
|
||||||
|
|
||||||
|
r.setdirty = function () {
|
||||||
|
r.dirty = true;
|
||||||
|
if (r.en) {
|
||||||
|
loadgrid();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
function setsz(v) {
|
||||||
|
if (v !== undefined) {
|
||||||
|
r.sz = v;
|
||||||
|
swrite('gridsz', r.sz);
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
document.documentElement.style.setProperty('--grid-sz', r.sz + 'em');
|
||||||
|
}
|
||||||
|
catch (ex) { }
|
||||||
|
}
|
||||||
|
setsz();
|
||||||
|
|
||||||
|
function seltgl(e) {
|
||||||
|
if (e && e.ctrlKey)
|
||||||
|
return true;
|
||||||
|
|
||||||
|
ev(e);
|
||||||
|
var oth = ebi(this.getAttribute('ref')),
|
||||||
|
td = oth.parentNode.nextSibling,
|
||||||
|
tr = td.parentNode;
|
||||||
|
|
||||||
|
td.click();
|
||||||
|
this.setAttribute('class', tr.getAttribute('class'));
|
||||||
|
}
|
||||||
|
|
||||||
|
function bgopen(e) {
|
||||||
|
ev(e);
|
||||||
|
var url = this.getAttribute('href');
|
||||||
|
window.open(url, '_blank');
|
||||||
|
}
|
||||||
|
|
||||||
|
r.loadsel = function () {
|
||||||
|
var ths = QSA('#ggrid>a'),
|
||||||
|
have_sel = !!QS('#files tr.sel');
|
||||||
|
|
||||||
|
for (var a = 0, aa = ths.length; a < aa; a++) {
|
||||||
|
ths[a].onclick = r.sel ? seltgl : have_sel ? bgopen : null;
|
||||||
|
ths[a].setAttribute('class', ebi(ths[a].getAttribute('ref')).parentNode.parentNode.getAttribute('class'));
|
||||||
|
}
|
||||||
|
var uns = QS('#ggrid a[ref="unsearch"]');
|
||||||
|
if (uns)
|
||||||
|
uns.onclick = function () {
|
||||||
|
ebi('unsearch').click();
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
function loadgrid() {
|
||||||
|
if (have_webp === null)
|
||||||
|
return setTimeout(loadgrid, 50);
|
||||||
|
|
||||||
|
if (!r.dirty)
|
||||||
|
return r.loadsel();
|
||||||
|
|
||||||
|
var html = [];
|
||||||
|
var files = QSA('#files>tbody>tr>td:nth-child(2) a[id]');
|
||||||
|
for (var a = 0, aa = files.length; a < aa; a++) {
|
||||||
|
var ao = files[a],
|
||||||
|
href = esc(ao.getAttribute('href')),
|
||||||
|
ref = ao.getAttribute('id'),
|
||||||
|
isdir = href.split('?')[0].slice(-1)[0] == '/',
|
||||||
|
ac = isdir ? ' class="dir"' : '',
|
||||||
|
ihref = href;
|
||||||
|
|
||||||
|
if (r.thumbs) {
|
||||||
|
ihref += (ihref.indexOf('?') === -1 ? '?' : '&') + 'th=' + (have_webp ? 'w' : 'j');
|
||||||
|
}
|
||||||
|
else if (isdir) {
|
||||||
|
ihref = '/.cpr/ico/folder';
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
var ar = href.split('?')[0].split('.');
|
||||||
|
if (ar.length > 1)
|
||||||
|
ar = ar.slice(1);
|
||||||
|
|
||||||
|
ihref = '';
|
||||||
|
ar.reverse();
|
||||||
|
for (var b = 0; b < ar.length; b++) {
|
||||||
|
if (ar[b].length > 7)
|
||||||
|
break;
|
||||||
|
|
||||||
|
ihref = ar[b] + '.' + ihref;
|
||||||
|
}
|
||||||
|
if (!ihref) {
|
||||||
|
ihref = 'unk.';
|
||||||
|
}
|
||||||
|
ihref = '/.cpr/ico/' + ihref.slice(0, -1);
|
||||||
|
}
|
||||||
|
|
||||||
|
html.push('<a href="' + href + '" ref="' + ref + '"><img src="' +
|
||||||
|
ihref + '" /><span' + ac + '>' + ao.innerHTML + '</span></a>');
|
||||||
|
}
|
||||||
|
lfiles.style.display = 'none';
|
||||||
|
gfiles.style.display = 'block';
|
||||||
|
ebi('ggrid').innerHTML = html.join('\n');
|
||||||
|
r.loadsel();
|
||||||
|
}
|
||||||
|
|
||||||
|
if (r.en) {
|
||||||
|
loadgrid();
|
||||||
|
}
|
||||||
|
|
||||||
|
return r;
|
||||||
|
})();
|
||||||
|
|
||||||
|
|
||||||
function tree_neigh(n) {
|
function tree_neigh(n) {
|
||||||
var links = QSA('#treeul li>a+a');
|
var links = QSA('#treeul li>a+a');
|
||||||
if (!links.length) {
|
if (!links.length) {
|
||||||
@@ -709,6 +969,23 @@ document.onkeydown = function (e) {
|
|||||||
|
|
||||||
if (k == 'KeyP')
|
if (k == 'KeyP')
|
||||||
return tree_up();
|
return tree_up();
|
||||||
|
|
||||||
|
if (k == 'KeyG')
|
||||||
|
return ebi('griden').click();
|
||||||
|
|
||||||
|
if (k == 'KeyT')
|
||||||
|
return ebi('thumbs').click();
|
||||||
|
|
||||||
|
if (thegrid.en) {
|
||||||
|
if (k == 'KeyS')
|
||||||
|
return ebi('gridsel').click();
|
||||||
|
|
||||||
|
if (k == 'KeyA')
|
||||||
|
return QSA('#ghead>a[z]')[0].click();
|
||||||
|
|
||||||
|
if (k == 'KeyD')
|
||||||
|
return QSA('#ghead>a[z]')[1].click();
|
||||||
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
@@ -765,6 +1042,7 @@ document.onkeydown = function (e) {
|
|||||||
for (var a = 0; a < trs.length; a += 2) {
|
for (var a = 0; a < trs.length; a += 2) {
|
||||||
html.push('<table>' + (trs[a].concat(trs[a + 1])).join('\n') + '</table>');
|
html.push('<table>' + (trs[a].concat(trs[a + 1])).join('\n') + '</table>');
|
||||||
}
|
}
|
||||||
|
html.push('<table id="tq_raw"><tr><td>raw</td><td><input id="q_raw" type="text" name="q" /></td></tr></table>');
|
||||||
ebi('srch_form').innerHTML = html.join('\n');
|
ebi('srch_form').innerHTML = html.join('\n');
|
||||||
|
|
||||||
var o = QSA('#op_search input');
|
var o = QSA('#op_search input');
|
||||||
@@ -789,33 +1067,83 @@ document.onkeydown = function (e) {
|
|||||||
var chk = ebi(id.slice(0, -1) + 'c');
|
var chk = ebi(id.slice(0, -1) + 'c');
|
||||||
chk.checked = ((v + '').length > 0);
|
chk.checked = ((v + '').length > 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (id != "q_raw")
|
||||||
|
encode_query();
|
||||||
|
|
||||||
clearTimeout(search_timeout);
|
clearTimeout(search_timeout);
|
||||||
if (Date.now() - search_in_progress > 30 * 1000)
|
if (Date.now() - search_in_progress > 30 * 1000)
|
||||||
search_timeout = setTimeout(do_search, 200);
|
search_timeout = setTimeout(do_search, 200);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function encode_query() {
|
||||||
|
var q = '';
|
||||||
|
for (var a = 0; a < sconf.length; a++) {
|
||||||
|
for (var b = 1; b < sconf[a].length; b++) {
|
||||||
|
var k = sconf[a][b][0],
|
||||||
|
chk = 'srch_' + k + 'c',
|
||||||
|
tvs = ebi('srch_' + k + 'v').value.split(/ /g);
|
||||||
|
|
||||||
|
if (!ebi(chk).checked)
|
||||||
|
continue;
|
||||||
|
|
||||||
|
for (var c = 0; c < tvs.length; c++) {
|
||||||
|
var tv = tvs[c];
|
||||||
|
if (!tv.length)
|
||||||
|
break;
|
||||||
|
|
||||||
|
q += ' and ';
|
||||||
|
|
||||||
|
if (k == 'adv') {
|
||||||
|
q += tv.replace(/ /g, " and ").replace(/([=!><]=?)/, " $1 ");
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (k.length == 3) {
|
||||||
|
q += k.replace(/sz/, 'size').replace(/dt/, 'date').replace(/l$/, ' >= ').replace(/u$/, ' <= ') + tv;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (k == 'path' || k == 'name' || k == 'tags') {
|
||||||
|
var not = ' ';
|
||||||
|
if (tv.slice(0, 1) == '-') {
|
||||||
|
tv = tv.slice(1);
|
||||||
|
not = ' not ';
|
||||||
|
}
|
||||||
|
|
||||||
|
if (tv.slice(0, 1) == '^') {
|
||||||
|
tv = tv.slice(1);
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
tv = '*' + tv;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (tv.slice(-1) == '$') {
|
||||||
|
tv = tv.slice(0, -1);
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
tv += '*';
|
||||||
|
}
|
||||||
|
|
||||||
|
q += k + not + 'like ' + tv;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ebi('q_raw').value = q.slice(5);
|
||||||
|
}
|
||||||
|
|
||||||
function do_search() {
|
function do_search() {
|
||||||
search_in_progress = Date.now();
|
search_in_progress = Date.now();
|
||||||
srch_msg(false, "searching...");
|
srch_msg(false, "searching...");
|
||||||
clearTimeout(search_timeout);
|
clearTimeout(search_timeout);
|
||||||
|
|
||||||
var params = {},
|
|
||||||
o = QSA('#op_search input[type="text"]');
|
|
||||||
|
|
||||||
for (var a = 0; a < o.length; a++) {
|
|
||||||
var chk = ebi(o[a].getAttribute('id').slice(0, -1) + 'c');
|
|
||||||
if (!chk.checked)
|
|
||||||
continue;
|
|
||||||
|
|
||||||
params[o[a].getAttribute('name')] = o[a].value;
|
|
||||||
}
|
|
||||||
// ebi('srch_q').textContent = JSON.stringify(params, null, 4);
|
|
||||||
var xhr = new XMLHttpRequest();
|
var xhr = new XMLHttpRequest();
|
||||||
xhr.open('POST', '/?srch', true);
|
xhr.open('POST', '/?srch', true);
|
||||||
xhr.setRequestHeader('Content-Type', 'text/plain');
|
xhr.setRequestHeader('Content-Type', 'text/plain');
|
||||||
xhr.onreadystatechange = xhr_search_results;
|
xhr.onreadystatechange = xhr_search_results;
|
||||||
xhr.ts = Date.now();
|
xhr.ts = Date.now();
|
||||||
xhr.send(JSON.stringify(params));
|
xhr.send(JSON.stringify({ "q": ebi('q_raw').value }));
|
||||||
}
|
}
|
||||||
|
|
||||||
function xhr_search_results() {
|
function xhr_search_results() {
|
||||||
@@ -864,7 +1192,7 @@ document.onkeydown = function (e) {
|
|||||||
sz = esc(r.sz + ''),
|
sz = esc(r.sz + ''),
|
||||||
rp = esc(r.rp + ''),
|
rp = esc(r.rp + ''),
|
||||||
ext = rp.lastIndexOf('.') > 0 ? rp.split('.').slice(-1)[0] : '%',
|
ext = rp.lastIndexOf('.') > 0 ? rp.split('.').slice(-1)[0] : '%',
|
||||||
links = linksplit(rp);
|
links = linksplit(r.rp + '');
|
||||||
|
|
||||||
if (ext.length > 8)
|
if (ext.length > 8)
|
||||||
ext = '%';
|
ext = '%';
|
||||||
@@ -928,7 +1256,6 @@ var treectl = (function () {
|
|||||||
treesz = icfg_get('treesz', 16);
|
treesz = icfg_get('treesz', 16);
|
||||||
|
|
||||||
treesz = Math.min(Math.max(treesz, 4), 50);
|
treesz = Math.min(Math.max(treesz, 4), 50);
|
||||||
console.log('treesz [' + treesz + ']');
|
|
||||||
|
|
||||||
function entree(e) {
|
function entree(e) {
|
||||||
ev(e);
|
ev(e);
|
||||||
@@ -987,7 +1314,7 @@ var treectl = (function () {
|
|||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
var top = Math.max(0, parseInt(wrap.offsetTop)),
|
var top = Math.max(0, parseInt(wrap.offsetTop)),
|
||||||
treeh = (winh - atop) - 4;
|
treeh = winh - atop;
|
||||||
|
|
||||||
tree.style.top = top + 'px';
|
tree.style.top = top + 'px';
|
||||||
tree.style.height = treeh < 10 ? '' : treeh + 'px';
|
tree.style.height = treeh < 10 ? '' : treeh + 'px';
|
||||||
@@ -1121,7 +1448,7 @@ var treectl = (function () {
|
|||||||
if (hpush)
|
if (hpush)
|
||||||
get_tree('.', xhr.top);
|
get_tree('.', xhr.top);
|
||||||
|
|
||||||
enspin('#files');
|
enspin(thegrid.en ? '#gfiles' : '#files');
|
||||||
}
|
}
|
||||||
|
|
||||||
function treegrow(e) {
|
function treegrow(e) {
|
||||||
@@ -1173,8 +1500,10 @@ var treectl = (function () {
|
|||||||
nodes = sortfiles(nodes);
|
nodes = sortfiles(nodes);
|
||||||
for (var a = 0; a < nodes.length; a++) {
|
for (var a = 0; a < nodes.length; a++) {
|
||||||
var r = nodes[a],
|
var r = nodes[a],
|
||||||
ln = ['<tr><td>' + r.lead + '</td><td><a href="' +
|
hname = esc(uricom_dec(r.href)[0]),
|
||||||
top + r.href + '">' + esc(uricom_dec(r.href)[0]) + '</a>', r.sz];
|
sortv = (r.href.slice(-1) == '/' ? '\t' : '') + hname,
|
||||||
|
ln = ['<tr><td>' + r.lead + '</td><td sortv="' + sortv +
|
||||||
|
'"><a href="' + top + r.href + '">' + hname + '</a>', r.sz];
|
||||||
|
|
||||||
for (var b = 0; b < res.taglist.length; b++) {
|
for (var b = 0; b < res.taglist.length; b++) {
|
||||||
var k = res.taglist[b],
|
var k = res.taglist[b],
|
||||||
@@ -1199,10 +1528,13 @@ var treectl = (function () {
|
|||||||
|
|
||||||
apply_perms(res.perms);
|
apply_perms(res.perms);
|
||||||
despin('#files');
|
despin('#files');
|
||||||
|
despin('#gfiles');
|
||||||
|
|
||||||
ebi('pro').innerHTML = res.logues ? res.logues[0] || "" : "";
|
ebi('pro').innerHTML = res.logues ? res.logues[0] || "" : "";
|
||||||
ebi('epi').innerHTML = res.logues ? res.logues[1] || "" : "";
|
ebi('epi').innerHTML = res.logues ? res.logues[1] || "" : "";
|
||||||
|
|
||||||
|
document.title = '⇆🎉 ' + uricom_dec(document.location.pathname.slice(1, -1))[0];
|
||||||
|
|
||||||
filecols.set_style();
|
filecols.set_style();
|
||||||
mukey.render();
|
mukey.render();
|
||||||
msel.render();
|
msel.render();
|
||||||
@@ -1338,7 +1670,7 @@ function apply_perms(perms) {
|
|||||||
up2k.set_fsearch();
|
up2k.set_fsearch();
|
||||||
|
|
||||||
ebi('widget').style.display = have_read ? '' : 'none';
|
ebi('widget').style.display = have_read ? '' : 'none';
|
||||||
ebi('files').style.display = have_read ? '' : 'none';
|
thegrid.setvis(have_read);
|
||||||
if (!have_read)
|
if (!have_read)
|
||||||
goto('up2k');
|
goto('up2k');
|
||||||
}
|
}
|
||||||
@@ -1731,6 +2063,7 @@ var msel = (function () {
|
|||||||
}
|
}
|
||||||
function selui() {
|
function selui() {
|
||||||
clmod(ebi('wtoggle'), 'sel', getsel().length);
|
clmod(ebi('wtoggle'), 'sel', getsel().length);
|
||||||
|
thegrid.loadsel();
|
||||||
}
|
}
|
||||||
function seltgl(e) {
|
function seltgl(e) {
|
||||||
ev(e);
|
ev(e);
|
||||||
@@ -1789,6 +2122,23 @@ var msel = (function () {
|
|||||||
})();
|
})();
|
||||||
|
|
||||||
|
|
||||||
|
(function () {
|
||||||
|
try {
|
||||||
|
var tr = ebi('files').tBodies[0].rows;
|
||||||
|
for (var a = 0; a < tr.length; a++) {
|
||||||
|
var td = tr[a].cells[1],
|
||||||
|
ao = td.firstChild,
|
||||||
|
href = ao.getAttribute('href'),
|
||||||
|
isdir = href.split('?')[0].slice(-1)[0] == '/',
|
||||||
|
txt = ao.textContent;
|
||||||
|
|
||||||
|
td.setAttribute('sortv', (isdir ? '\t' : '') + txt);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
catch (ex) { }
|
||||||
|
})();
|
||||||
|
|
||||||
|
|
||||||
function ev_row_tgl(e) {
|
function ev_row_tgl(e) {
|
||||||
ev(e);
|
ev(e);
|
||||||
filecols.toggle(this.parentElement.parentElement.getElementsByTagName('span')[0].textContent);
|
filecols.toggle(this.parentElement.parentElement.getElementsByTagName('span')[0].textContent);
|
||||||
@@ -1839,6 +2189,8 @@ function reload_browser(not_mp) {
|
|||||||
|
|
||||||
if (window['up2k'])
|
if (window['up2k'])
|
||||||
up2k.set_fsearch();
|
up2k.set_fsearch();
|
||||||
|
|
||||||
|
thegrid.setdirty();
|
||||||
}
|
}
|
||||||
reload_browser(true);
|
reload_browser(true);
|
||||||
mukey.render();
|
mukey.render();
|
||||||
|
|||||||
Binary file not shown.
|
Before Width: | Height: | Size: 248 B |
@@ -26,6 +26,26 @@ a {
|
|||||||
border-radius: .2em;
|
border-radius: .2em;
|
||||||
padding: .2em .8em;
|
padding: .2em .8em;
|
||||||
}
|
}
|
||||||
|
table {
|
||||||
|
border-collapse: collapse;
|
||||||
|
}
|
||||||
|
.vols td,
|
||||||
|
.vols th {
|
||||||
|
padding: .3em .6em;
|
||||||
|
text-align: left;
|
||||||
|
}
|
||||||
|
.num {
|
||||||
|
border-right: 1px solid #bbb;
|
||||||
|
}
|
||||||
|
.num td {
|
||||||
|
padding: .1em .7em .1em 0;
|
||||||
|
}
|
||||||
|
.num td:first-child {
|
||||||
|
text-align: right;
|
||||||
|
}
|
||||||
|
.btns {
|
||||||
|
margin: 1em 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
html.dark,
|
html.dark,
|
||||||
@@ -51,3 +71,6 @@ html.dark input {
|
|||||||
padding: .5em .7em;
|
padding: .5em .7em;
|
||||||
margin: 0 .5em 0 0;
|
margin: 0 .5em 0 0;
|
||||||
}
|
}
|
||||||
|
html.dark .num {
|
||||||
|
border-color: #777;
|
||||||
|
}
|
||||||
@@ -13,11 +13,37 @@
|
|||||||
<div id="wrap">
|
<div id="wrap">
|
||||||
<p>hello {{ this.uname }}</p>
|
<p>hello {{ this.uname }}</p>
|
||||||
|
|
||||||
|
{%- if avol %}
|
||||||
|
<h1>admin panel:</h1>
|
||||||
|
<table><tr><td> <!-- hehehe -->
|
||||||
|
<table class="num">
|
||||||
|
<tr><td>scanning</td><td>{{ scanning }}</td></tr>
|
||||||
|
<tr><td>hash-q</td><td>{{ hashq }}</td></tr>
|
||||||
|
<tr><td>tag-q</td><td>{{ tagq }}</td></tr>
|
||||||
|
<tr><td>mtp-q</td><td>{{ mtpq }}</td></tr>
|
||||||
|
</table>
|
||||||
|
</td><td>
|
||||||
|
<table class="vols">
|
||||||
|
<thead><tr><th>vol</th><th>action</th><th>status</th></tr></thead>
|
||||||
|
<tbody>
|
||||||
|
{% for mp in avol %}
|
||||||
|
{%- if mp in vstate and vstate[mp] %}
|
||||||
|
<tr><td><a href="{{ mp }}{{ url_suf }}">{{ mp }}</a></td><td><a href="{{ mp }}?scan">rescan</a></td><td>{{ vstate[mp] }}</td></tr>
|
||||||
|
{%- endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</tbody>
|
||||||
|
</table>
|
||||||
|
</td></tr></table>
|
||||||
|
<div class="btns">
|
||||||
|
<a href="{{ avol[0] }}?stack">dump stack</a>
|
||||||
|
</div>
|
||||||
|
{%- endif %}
|
||||||
|
|
||||||
{%- if rvol %}
|
{%- if rvol %}
|
||||||
<h1>you can browse these:</h1>
|
<h1>you can browse these:</h1>
|
||||||
<ul>
|
<ul>
|
||||||
{% for mp in rvol %}
|
{% for mp in rvol %}
|
||||||
<li><a href="/{{ mp }}{{ url_suf }}">/{{ mp }}</a></li>
|
<li><a href="{{ mp }}{{ url_suf }}">{{ mp }}</a></li>
|
||||||
{% endfor %}
|
{% endfor %}
|
||||||
</ul>
|
</ul>
|
||||||
{%- endif %}
|
{%- endif %}
|
||||||
@@ -26,7 +52,7 @@
|
|||||||
<h1>you can upload to:</h1>
|
<h1>you can upload to:</h1>
|
||||||
<ul>
|
<ul>
|
||||||
{% for mp in wvol %}
|
{% for mp in wvol %}
|
||||||
<li><a href="/{{ mp }}{{ url_suf }}">/{{ mp }}</a></li>
|
<li><a href="{{ mp }}{{ url_suf }}">{{ mp }}</a></li>
|
||||||
{% endfor %}
|
{% endfor %}
|
||||||
</ul>
|
</ul>
|
||||||
{%- endif %}
|
{%- endif %}
|
||||||
|
|||||||
@@ -17,14 +17,16 @@ function goto_up2k() {
|
|||||||
// chrome requires https to use crypto.subtle,
|
// chrome requires https to use crypto.subtle,
|
||||||
// usually it's undefined but some chromes throw on invoke
|
// usually it's undefined but some chromes throw on invoke
|
||||||
var up2k = null;
|
var up2k = null;
|
||||||
|
var sha_js = window.WebAssembly ? 'hw' : 'ac'; // ff53,c57,sa11
|
||||||
try {
|
try {
|
||||||
var cf = crypto.subtle || crypto.webkitSubtle;
|
var cf = crypto.subtle || crypto.webkitSubtle;
|
||||||
cf.digest('SHA-512', new Uint8Array(1)).then(
|
cf.digest('SHA-512', new Uint8Array(1)).then(
|
||||||
function (x) { up2k = up2k_init(cf) },
|
function (x) { console.log('sha-ok'); up2k = up2k_init(cf); },
|
||||||
function (x) { up2k = up2k_init(false) }
|
function (x) { console.log('sha-ng:', x); up2k = up2k_init(false); }
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
catch (ex) {
|
catch (ex) {
|
||||||
|
console.log('sha-na:', ex);
|
||||||
try {
|
try {
|
||||||
up2k = up2k_init(false);
|
up2k = up2k_init(false);
|
||||||
}
|
}
|
||||||
@@ -429,13 +431,15 @@ function up2k_init(subtle) {
|
|||||||
// upload ui hidden by default, clicking the header shows it
|
// upload ui hidden by default, clicking the header shows it
|
||||||
function init_deps() {
|
function init_deps() {
|
||||||
if (!subtle && !window.asmCrypto) {
|
if (!subtle && !window.asmCrypto) {
|
||||||
showmodal('<h1>loading sha512.js</h1><h2>since ' + shame + '</h2><h4>thanks chrome</h4>');
|
var fn = 'sha512.' + sha_js + '.js';
|
||||||
import_js('/.cpr/deps/sha512.js', unmodal);
|
showmodal('<h1>loading ' + fn + '</h1><h2>since ' + shame + '</h2><h4>thanks chrome</h4>');
|
||||||
|
import_js('/.cpr/deps/' + fn, unmodal);
|
||||||
|
|
||||||
if (is_https)
|
if (is_https)
|
||||||
ebi('u2foot').innerHTML = shame + ' so <em>this</em> uploader will do like 500kB/s at best';
|
ebi('u2foot').innerHTML = shame + ' so <em>this</em> uploader will do like 500kB/s at best';
|
||||||
else
|
else
|
||||||
ebi('u2foot').innerHTML = 'seems like ' + shame + ' so do that if you want more performance';
|
ebi('u2foot').innerHTML = 'seems like ' + shame + ' so do that if you want more performance <span style="color:#' +
|
||||||
|
(sha_js == 'ac' ? 'c84">(expecting 20' : '8a5">(but dont worry too much, expect 100') + ' MiB/s)</span>';
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -695,7 +699,7 @@ function up2k_init(subtle) {
|
|||||||
|
|
||||||
pvis.addfile([
|
pvis.addfile([
|
||||||
fsearch ? esc(entry.name) : linksplit(
|
fsearch ? esc(entry.name) : linksplit(
|
||||||
esc(uricom_dec(entry.purl)[0] + entry.name)).join(' '),
|
uricom_dec(entry.purl)[0] + entry.name).join(' '),
|
||||||
'📐 hash',
|
'📐 hash',
|
||||||
''
|
''
|
||||||
], fobj.size);
|
], fobj.size);
|
||||||
@@ -885,6 +889,10 @@ function up2k_init(subtle) {
|
|||||||
return base64;
|
return base64;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function hex2u8(txt) {
|
||||||
|
return new Uint8Array(txt.match(/.{2}/g).map(function (b) { return parseInt(b, 16); }));
|
||||||
|
}
|
||||||
|
|
||||||
function get_chunksize(filesize) {
|
function get_chunksize(filesize) {
|
||||||
var chunksize = 1024 * 1024,
|
var chunksize = 1024 * 1024,
|
||||||
stepsize = 512 * 1024;
|
stepsize = 512 * 1024;
|
||||||
@@ -986,10 +994,18 @@ function up2k_init(subtle) {
|
|||||||
if (subtle)
|
if (subtle)
|
||||||
subtle.digest('SHA-512', buf).then(hash_done);
|
subtle.digest('SHA-512', buf).then(hash_done);
|
||||||
else setTimeout(function () {
|
else setTimeout(function () {
|
||||||
|
var u8buf = new Uint8Array(buf);
|
||||||
|
if (sha_js == 'hw') {
|
||||||
|
hashwasm.sha512(u8buf).then(function (v) {
|
||||||
|
hash_done(hex2u8(v))
|
||||||
|
});
|
||||||
|
}
|
||||||
|
else {
|
||||||
var hasher = new asmCrypto.Sha512();
|
var hasher = new asmCrypto.Sha512();
|
||||||
hasher.process(new Uint8Array(buf));
|
hasher.process(u8buf);
|
||||||
hasher.finish();
|
hasher.finish();
|
||||||
hash_done(hasher.result);
|
hash_done(hasher.result);
|
||||||
|
}
|
||||||
}, 1);
|
}, 1);
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -1022,7 +1038,7 @@ function up2k_init(subtle) {
|
|||||||
else {
|
else {
|
||||||
smsg = 'found';
|
smsg = 'found';
|
||||||
var hit = response.hits[0],
|
var hit = response.hits[0],
|
||||||
msg = linksplit(esc(hit.rp)).join(''),
|
msg = linksplit(hit.rp).join(''),
|
||||||
tr = unix2iso(hit.ts),
|
tr = unix2iso(hit.ts),
|
||||||
tu = unix2iso(t.lmod),
|
tu = unix2iso(t.lmod),
|
||||||
diff = parseInt(t.lmod) - parseInt(hit.ts),
|
diff = parseInt(t.lmod) - parseInt(hit.ts),
|
||||||
@@ -1044,7 +1060,7 @@ function up2k_init(subtle) {
|
|||||||
if (response.name !== t.name) {
|
if (response.name !== t.name) {
|
||||||
// file exists; server renamed us
|
// file exists; server renamed us
|
||||||
t.name = response.name;
|
t.name = response.name;
|
||||||
pvis.seth(t.n, 0, linksplit(esc(t.purl + t.name)).join(' '));
|
pvis.seth(t.n, 0, linksplit(t.purl + t.name).join(' '));
|
||||||
}
|
}
|
||||||
|
|
||||||
var chunksize = get_chunksize(t.size),
|
var chunksize = get_chunksize(t.size),
|
||||||
|
|||||||
@@ -238,6 +238,10 @@
|
|||||||
color: #fff;
|
color: #fff;
|
||||||
font-style: italic;
|
font-style: italic;
|
||||||
}
|
}
|
||||||
|
#u2foot span {
|
||||||
|
color: #999;
|
||||||
|
font-size: .9em;
|
||||||
|
}
|
||||||
#u2footfoot {
|
#u2footfoot {
|
||||||
margin-bottom: -1em;
|
margin-bottom: -1em;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -135,7 +135,7 @@ function clmod(obj, cls, add) {
|
|||||||
|
|
||||||
|
|
||||||
function sortfiles(nodes) {
|
function sortfiles(nodes) {
|
||||||
var sopts = jread('fsort', [["lead", -1, ""], ["href", 1, ""]]);
|
var sopts = jread('fsort', [["href", 1, ""]]);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
var is_srch = false;
|
var is_srch = false;
|
||||||
@@ -152,6 +152,9 @@ function sortfiles(nodes) {
|
|||||||
if (!name)
|
if (!name)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
if (name == 'ts')
|
||||||
|
typ = 'int';
|
||||||
|
|
||||||
if (name.indexOf('tags/') === 0) {
|
if (name.indexOf('tags/') === 0) {
|
||||||
name = name.slice(5);
|
name = name.slice(5);
|
||||||
for (var b = 0, bb = nodes.length; b < bb; b++)
|
for (var b = 0, bb = nodes.length; b < bb; b++)
|
||||||
@@ -163,8 +166,12 @@ function sortfiles(nodes) {
|
|||||||
|
|
||||||
if ((v + '').indexOf('<a ') === 0)
|
if ((v + '').indexOf('<a ') === 0)
|
||||||
v = v.split('>')[1];
|
v = v.split('>')[1];
|
||||||
else if (name == "href" && v)
|
else if (name == "href" && v) {
|
||||||
|
if (v.slice(-1) == '/')
|
||||||
|
v = '\t' + v;
|
||||||
|
|
||||||
v = uricom_dec(v)[0]
|
v = uricom_dec(v)[0]
|
||||||
|
}
|
||||||
|
|
||||||
nodes[b]._sv = v;
|
nodes[b]._sv = v;
|
||||||
}
|
}
|
||||||
@@ -198,6 +205,8 @@ function sortfiles(nodes) {
|
|||||||
}
|
}
|
||||||
catch (ex) {
|
catch (ex) {
|
||||||
console.log("failed to apply sort config: " + ex);
|
console.log("failed to apply sort config: " + ex);
|
||||||
|
console.log("resetting fsort " + sread('fsort'))
|
||||||
|
localStorage.removeItem('fsort');
|
||||||
}
|
}
|
||||||
return nodes;
|
return nodes;
|
||||||
}
|
}
|
||||||
@@ -349,12 +358,16 @@ function linksplit(rp) {
|
|||||||
link = rp.slice(0, ofs + 1);
|
link = rp.slice(0, ofs + 1);
|
||||||
rp = rp.slice(ofs + 1);
|
rp = rp.slice(ofs + 1);
|
||||||
}
|
}
|
||||||
var vlink = link;
|
var vlink = esc(link),
|
||||||
if (link.indexOf('/') !== -1)
|
elink = uricom_enc(link);
|
||||||
vlink = link.slice(0, -1) + '<span>/</span>';
|
|
||||||
|
|
||||||
ret.push('<a href="' + apath + link + '">' + vlink + '</a>');
|
if (link.indexOf('/') !== -1) {
|
||||||
apath += link;
|
vlink = vlink.slice(0, -1) + '<span>/</span>';
|
||||||
|
elink = elink.slice(0, -3) + '/';
|
||||||
|
}
|
||||||
|
|
||||||
|
ret.push('<a href="' + apath + elink + '">' + vlink + '</a>');
|
||||||
|
apath += elink;
|
||||||
}
|
}
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
@@ -456,11 +469,15 @@ function jwrite(key, val) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
function icfg_get(name, defval) {
|
function icfg_get(name, defval) {
|
||||||
|
return parseInt(fcfg_get(name, defval));
|
||||||
|
}
|
||||||
|
|
||||||
|
function fcfg_get(name, defval) {
|
||||||
var o = ebi(name);
|
var o = ebi(name);
|
||||||
|
|
||||||
var val = parseInt(sread(name));
|
var val = parseFloat(sread(name));
|
||||||
if (isNaN(val))
|
if (isNaN(val))
|
||||||
return parseInt(o ? o.value : defval);
|
return parseFloat(o ? o.value : defval);
|
||||||
|
|
||||||
if (o)
|
if (o)
|
||||||
o.value = val;
|
o.value = val;
|
||||||
|
|||||||
11
docs/README.md
Normal file
11
docs/README.md
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
## [`minimal-up2k.html`](minimal-up2k.html)
|
||||||
|
* save as `.epilogue.html` inside a folder to [simplify the ui](https://user-images.githubusercontent.com/241032/118311195-dd6ca380-b4ef-11eb-86f3-75a3ff2e1332.png)
|
||||||
|
|
||||||
|
## [`browser.css`](browser.css)
|
||||||
|
* example for `--css-browser`
|
||||||
|
|
||||||
|
## [`rclone.md`](rclone.md)
|
||||||
|
* notes on using rclone as a fuse client/server
|
||||||
|
|
||||||
|
## [`example.conf`](example.conf)
|
||||||
|
* example config file for `-c` which never really happened
|
||||||
29
docs/browser.css
Normal file
29
docs/browser.css
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
html {
|
||||||
|
background: url('/wp/wallhaven-mdjrqy.jpg') center / cover no-repeat fixed;
|
||||||
|
}
|
||||||
|
#files th {
|
||||||
|
background: rgba(32, 32, 32, 0.9) !important;
|
||||||
|
}
|
||||||
|
#ops,
|
||||||
|
#treeul,
|
||||||
|
#files td {
|
||||||
|
background: rgba(32, 32, 32, 0.3) !important;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
html.light {
|
||||||
|
background: url('/wp/wallhaven-dpxl6l.png') center / cover no-repeat fixed;
|
||||||
|
}
|
||||||
|
html.light #files th {
|
||||||
|
background: rgba(255, 255, 255, 0.9) !important;
|
||||||
|
}
|
||||||
|
html.light #ops,
|
||||||
|
html.light #treeul,
|
||||||
|
html.light #files td {
|
||||||
|
background: rgba(248, 248, 248, 0.8) !important;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#files * {
|
||||||
|
background: transparent !important;
|
||||||
|
}
|
||||||
@@ -7,7 +7,7 @@
|
|||||||
|
|
||||||
/* make the up2k ui REALLY minimal by hiding a bunch of stuff: */
|
/* make the up2k ui REALLY minimal by hiding a bunch of stuff: */
|
||||||
|
|
||||||
#ops, #tree, #path, #wrap>h2:last-child /* main tabs and navigators (tree/breadcrumbs) */
|
#ops, #tree, #path, #wrap>h2:last-child, /* main tabs and navigators (tree/breadcrumbs) */
|
||||||
|
|
||||||
#u2cleanup, #u2conf tr:first-child>td[rowspan]:not(#u2btn_cw), /* most of the config options */
|
#u2cleanup, #u2conf tr:first-child>td[rowspan]:not(#u2btn_cw), /* most of the config options */
|
||||||
|
|
||||||
|
|||||||
@@ -80,6 +80,13 @@ command -v gdate && date() { gdate "$@"; }; while true; do t=$(date +%s.%N); (ti
|
|||||||
var t=[]; var b=document.location.href.split('#')[0].slice(0, -1); document.querySelectorAll('#u2tab .prog a').forEach((x) => {t.push(b+encodeURI(x.getAttribute("href")))}); console.log(t.join("\n"));
|
var t=[]; var b=document.location.href.split('#')[0].slice(0, -1); document.querySelectorAll('#u2tab .prog a').forEach((x) => {t.push(b+encodeURI(x.getAttribute("href")))}); console.log(t.join("\n"));
|
||||||
|
|
||||||
|
|
||||||
|
##
|
||||||
|
## bash oneliners
|
||||||
|
|
||||||
|
# get the size and video-id of all youtube vids in folder, assuming filename ends with -id.ext, and create a copyparty search query
|
||||||
|
find -maxdepth 1 -printf '%s %p\n' | sort -n | awk '!/-([0-9a-zA-Z_-]{11})\.(mkv|mp4|webm)$/{next} {sub(/\.[^\.]+$/,"");n=length($0);v=substr($0,n-10);print $1, v}' | tee /dev/stderr | awk 'BEGIN {p="("} {printf("%s name like *-%s.* ",p,$2);p="or"} END {print ")\n"}' | cat >&2
|
||||||
|
|
||||||
|
|
||||||
##
|
##
|
||||||
## sqlite3 stuff
|
## sqlite3 stuff
|
||||||
|
|
||||||
@@ -146,6 +153,9 @@ dbg.asyncStore.pendingBreakpoints = {}
|
|||||||
# fix firefox phantom breakpoints
|
# fix firefox phantom breakpoints
|
||||||
about:config >> devtools.debugger.prefs-schema-version = -1
|
about:config >> devtools.debugger.prefs-schema-version = -1
|
||||||
|
|
||||||
|
# determine server version
|
||||||
|
git reset --hard origin/HEAD && git log --format=format:"%H %ai %d" --decorate=full > /dev/shm/revs && cat /dev/shm/revs | while read -r rev extra; do (git reset --hard $rev >/dev/null 2>/dev/null && dsz=$(cat copyparty/web/{util,browser,up2k}.js 2>/dev/null | diff -wNarU0 - <(cat /mnt/Users/ed/Downloads/ref/{util,browser,up2k}.js) | wc -c) && printf '%s %6s %s\n' "$rev" $dsz "$extra") </dev/null; done
|
||||||
|
|
||||||
|
|
||||||
##
|
##
|
||||||
## http 206
|
## http 206
|
||||||
|
|||||||
82
docs/nuitka.txt
Normal file
82
docs/nuitka.txt
Normal file
@@ -0,0 +1,82 @@
|
|||||||
|
# recipe for building an exe with nuitka (extreme jank edition)
|
||||||
|
#
|
||||||
|
# NOTE: win7 and win10 builds both work on win10 but
|
||||||
|
# on win7 they immediately c0000005 in kernelbase.dll
|
||||||
|
#
|
||||||
|
# first install python-3.6.8-amd64.exe
|
||||||
|
# [x] add to path
|
||||||
|
#
|
||||||
|
# copypaste the rest of this file into cmd
|
||||||
|
|
||||||
|
rem from pypi
|
||||||
|
cd \users\ed\downloads
|
||||||
|
python -m pip install --user Nuitka-0.6.14.7.tar.gz
|
||||||
|
|
||||||
|
rem https://github.com/brechtsanders/winlibs_mingw/releases/download/10.2.0-11.0.0-8.0.0-r5/winlibs-x86_64-posix-seh-gcc-10.2.0-llvm-11.0.0-mingw-w64-8.0.0-r5.zip
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\gcc\
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\gcc\x86_64\
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\gcc\x86_64\10.2.0-11.0.0-8.0.0-r5\
|
||||||
|
copy c:\users\ed\downloads\winlibs-x86_64-posix-seh-gcc-10.2.0-llvm-11.0.0-mingw-w64-8.0.0-r5.zip C:\Users\ed\AppData\Local\Nuitka\Nuitka\gcc\x86_64\10.2.0-11.0.0-8.0.0-r5\winlibs-x86_64-posix-seh-gcc-10.2.0-llvm-11.0.0-mingw-w64-8.0.0-r5.zip
|
||||||
|
|
||||||
|
rem https://github.com/ccache/ccache/releases/download/v3.7.12/ccache-3.7.12-windows-32.zip
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\ccache\
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\ccache\v3.7.12\
|
||||||
|
copy c:\users\ed\downloads\ccache-3.7.12-windows-32.zip C:\Users\ed\AppData\Local\Nuitka\Nuitka\ccache\v3.7.12\ccache-3.7.12-windows-32.zip
|
||||||
|
|
||||||
|
rem https://dependencywalker.com/depends22_x64.zip
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\depends\
|
||||||
|
mkdir C:\Users\ed\AppData\Local\Nuitka\Nuitka\depends\x86_64\
|
||||||
|
copy c:\users\ed\downloads\depends22_x64.zip C:\Users\ed\AppData\Local\Nuitka\Nuitka\depends\x86_64\depends22_x64.zip
|
||||||
|
|
||||||
|
cd \
|
||||||
|
rd /s /q %appdata%\..\local\temp\pe-copyparty
|
||||||
|
cd \users\ed\downloads
|
||||||
|
python copyparty-sfx.py -h
|
||||||
|
cd %appdata%\..\local\temp\pe-copyparty\copyparty
|
||||||
|
|
||||||
|
python
|
||||||
|
import os, re
|
||||||
|
os.rename('../dep-j2/jinja2', '../jinja2')
|
||||||
|
os.rename('../dep-j2/markupsafe', '../markupsafe')
|
||||||
|
|
||||||
|
print("# nuitka dies if .__init__.stuff is imported")
|
||||||
|
with open('__init__.py','r',encoding='utf-8') as f:
|
||||||
|
t1 = f.read()
|
||||||
|
|
||||||
|
with open('util.py','r',encoding='utf-8') as f:
|
||||||
|
t2 = f.read().split('\n')[3:]
|
||||||
|
|
||||||
|
t2 = [x for x in t2 if 'from .__init__' not in x]
|
||||||
|
t = t1 + '\n'.join(t2)
|
||||||
|
with open('__init__.py','w',encoding='utf-8') as f:
|
||||||
|
f.write('\n')
|
||||||
|
|
||||||
|
with open('util.py','w',encoding='utf-8') as f:
|
||||||
|
f.write(t)
|
||||||
|
|
||||||
|
print("# local-imports fail, prefix module names")
|
||||||
|
ptn = re.compile(r'^( *from )(\.[^ ]+ import .*)')
|
||||||
|
for d, _, fs in os.walk('.'):
|
||||||
|
for f in fs:
|
||||||
|
fp = os.path.join(d, f)
|
||||||
|
if not fp.endswith('.py'):
|
||||||
|
continue
|
||||||
|
t = ''
|
||||||
|
with open(fp,'r',encoding='utf-8') as f:
|
||||||
|
for ln in [x.rstrip('\r\n') for x in f]:
|
||||||
|
m = ptn.match(ln)
|
||||||
|
if not m:
|
||||||
|
t += ln + '\n'
|
||||||
|
continue
|
||||||
|
p1, p2 = m.groups()
|
||||||
|
t += "{}copyparty{}\n".format(p1, p2).replace("__init__", "util")
|
||||||
|
with open(fp,'w',encoding='utf-8') as f:
|
||||||
|
f.write(t)
|
||||||
|
|
||||||
|
exit()
|
||||||
|
|
||||||
|
cd ..
|
||||||
|
|
||||||
|
rd /s /q bout & python -m nuitka --standalone --onefile --windows-onefile-tempdir --python-flag=no_site --assume-yes-for-downloads --include-data-dir=copyparty\web=copyparty\web --include-data-dir=copyparty\res=copyparty\res --run --output-dir=bout --mingw64 --include-package=markupsafe --include-package=jinja2 copyparty
|
||||||
@@ -9,6 +9,12 @@ ENV ver_asmcrypto=5b994303a9d3e27e0915f72a10b6c2c51535a4dc \
|
|||||||
ver_zopfli=1.0.3
|
ver_zopfli=1.0.3
|
||||||
|
|
||||||
|
|
||||||
|
# TODO
|
||||||
|
# sha512.hw.js https://github.com/Daninet/hash-wasm
|
||||||
|
# sha512.kc.js https://github.com/chm-diederichs/sha3-wasm
|
||||||
|
# awk '/HMAC state/{o=1} /var HEAP/{o=0} /function hmac_reset/{o=1} /return \{/{o=0} /var __extends =/{o=1} /var Hash =/{o=0} /hmac_|pbkdf2_/{next} o{next} {gsub(/IllegalStateError/,"Exception")} {sub(/^ +/,"");sub(/^\/\/ .*/,"");sub(/;$/," ;")} 1' <sha512.ac.js.orig >sha512.ac.js; for fn in sha512.ac.js.orig sha512.ac.js; do wc -c <$fn; wc -c <$fn.gz ; for n in {1..9}; do printf '%8d %d bz\n' $(bzip2 -c$n <$fn | wc -c) $n; done; done
|
||||||
|
|
||||||
|
|
||||||
# download;
|
# download;
|
||||||
# the scp url is latin from https://fonts.googleapis.com/css2?family=Source+Code+Pro&display=swap
|
# the scp url is latin from https://fonts.googleapis.com/css2?family=Source+Code+Pro&display=swap
|
||||||
RUN mkdir -p /z/dist/no-pk \
|
RUN mkdir -p /z/dist/no-pk \
|
||||||
|
|||||||
12
scripts/install-githooks.sh
Executable file
12
scripts/install-githooks.sh
Executable file
@@ -0,0 +1,12 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
[ -e setup.py ] || ..
|
||||||
|
[ -e setup.py ] || {
|
||||||
|
echo u wot
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
|
||||||
|
cd .git/hooks
|
||||||
|
rm -f pre-commit
|
||||||
|
ln -s ../../scripts/run-tests.sh pre-commit
|
||||||
@@ -32,6 +32,10 @@ gtar=$(command -v gtar || command -v gnutar) || true
|
|||||||
[ -e /opt/local/bin/bzip2 ] &&
|
[ -e /opt/local/bin/bzip2 ] &&
|
||||||
bzip2() { /opt/local/bin/bzip2 "$@"; }
|
bzip2() { /opt/local/bin/bzip2 "$@"; }
|
||||||
}
|
}
|
||||||
|
|
||||||
|
gawk=$(command -v gawk || command -v gnuawk || command -v awk)
|
||||||
|
awk() { $gawk "$@"; }
|
||||||
|
|
||||||
pybin=$(command -v python3 || command -v python) || {
|
pybin=$(command -v python3 || command -v python) || {
|
||||||
echo need python
|
echo need python
|
||||||
exit 1
|
exit 1
|
||||||
@@ -194,17 +198,46 @@ tmv "$f"
|
|||||||
|
|
||||||
# up2k goes from 28k to 22k laff
|
# up2k goes from 28k to 22k laff
|
||||||
echo entabbening
|
echo entabbening
|
||||||
find | grep -E '\.(js|css|html)$' | while IFS= read -r f; do
|
find | grep -E '\.css$' | while IFS= read -r f; do
|
||||||
|
awk '{
|
||||||
|
sub(/^[ \t]+/,"");
|
||||||
|
sub(/[ \t]+$/,"");
|
||||||
|
$0=gensub(/^([a-z-]+) *: *(.*[^ ]) *;$/,"\\1:\\2;","1");
|
||||||
|
sub(/ +\{$/,"{");
|
||||||
|
gsub(/, /,",")
|
||||||
|
}
|
||||||
|
!/\}$/ {printf "%s",$0;next}
|
||||||
|
1
|
||||||
|
' <$f | gsed 's/;\}$/}/' >t
|
||||||
|
tmv "$f"
|
||||||
|
done
|
||||||
|
find | grep -E '\.(js|html)$' | while IFS= read -r f; do
|
||||||
unexpand -t 4 --first-only <"$f" >t
|
unexpand -t 4 --first-only <"$f" >t
|
||||||
tmv "$f"
|
tmv "$f"
|
||||||
done
|
done
|
||||||
|
|
||||||
|
|
||||||
|
gzres() {
|
||||||
|
command -v pigz &&
|
||||||
|
pk='pigz -11 -J 34 -I 100' ||
|
||||||
|
pk='gzip'
|
||||||
|
|
||||||
|
echo "$pk"
|
||||||
|
find | grep -E '\.(js|css)$' | while IFS= read -r f; do
|
||||||
|
echo -n .
|
||||||
|
$pk "$f"
|
||||||
|
done
|
||||||
|
echo
|
||||||
|
}
|
||||||
|
gzres
|
||||||
|
|
||||||
|
|
||||||
echo gen tarlist
|
echo gen tarlist
|
||||||
for d in copyparty dep-j2; do find $d -type f; done |
|
for d in copyparty dep-j2; do find $d -type f; done |
|
||||||
sed -r 's/(.*)\.(.*)/\2 \1/' | LC_ALL=C sort |
|
sed -r 's/(.*)\.(.*)/\2 \1/' | LC_ALL=C sort |
|
||||||
sed -r 's/([^ ]*) (.*)/\2.\1/' | grep -vE '/list1?$' > list1
|
sed -r 's/([^ ]*) (.*)/\2.\1/' | grep -vE '/list1?$' > list1
|
||||||
|
|
||||||
(grep -vE 'gz$' list1; grep -E 'gz$' list1) >list
|
(grep -vE '\.(gz|br)$' list1; grep -E '\.(gz|br)$' list1) >list || true
|
||||||
|
|
||||||
echo creating tar
|
echo creating tar
|
||||||
args=(--owner=1000 --group=1000)
|
args=(--owner=1000 --group=1000)
|
||||||
|
|||||||
34
scripts/profile.py
Normal file
34
scripts/profile.py
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
import sys
|
||||||
|
|
||||||
|
sys.path.insert(0, ".")
|
||||||
|
cmd = sys.argv[1]
|
||||||
|
|
||||||
|
if cmd == "cpp":
|
||||||
|
from copyparty.__main__ import main
|
||||||
|
|
||||||
|
argv = ["__main__", "-v", "srv::r", "-v", "../../yt:yt:r"]
|
||||||
|
main(argv=argv)
|
||||||
|
|
||||||
|
elif cmd == "test":
|
||||||
|
from unittest import main
|
||||||
|
|
||||||
|
argv = ["__main__", "discover", "-s", "tests"]
|
||||||
|
main(module=None, argv=argv)
|
||||||
|
|
||||||
|
else:
|
||||||
|
raise Exception()
|
||||||
|
|
||||||
|
# import dis; print(dis.dis(main))
|
||||||
|
|
||||||
|
|
||||||
|
# macos:
|
||||||
|
# option1) python3.9 -m pip install --user -U vmprof==0.4.9
|
||||||
|
# option2) python3.9 -m pip install --user -U https://github.com/vmprof/vmprof-python/archive/refs/heads/master.zip
|
||||||
|
#
|
||||||
|
# python -m vmprof -o prof --lines ./scripts/profile.py test
|
||||||
|
|
||||||
|
# linux: ~/.local/bin/vmprofshow prof tree | grep -vF '[1m 0.'
|
||||||
|
# macos: ~/Library/Python/3.9/bin/vmprofshow prof tree | grep -vF '[1m 0.'
|
||||||
|
# win: %appdata%\..\Roaming\Python\Python39\Scripts\vmprofshow.exe prof tree
|
||||||
15
scripts/run-tests.sh
Executable file
15
scripts/run-tests.sh
Executable file
@@ -0,0 +1,15 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
pids=()
|
||||||
|
for py in python{2,3}; do
|
||||||
|
nice $py -m unittest discover -s tests >/dev/null &
|
||||||
|
pids+=($!)
|
||||||
|
done
|
||||||
|
|
||||||
|
python3 scripts/test/smoketest.py &
|
||||||
|
pids+=($!)
|
||||||
|
|
||||||
|
for pid in ${pids[@]}; do
|
||||||
|
wait $pid
|
||||||
|
done
|
||||||
@@ -47,7 +47,7 @@ grep -E '/(python|pypy)[0-9\.-]*$' >$dir/pys || true
|
|||||||
printf '\033[1;30mlooking for jinja2 in [%s]\033[0m\n' "$_py" >&2
|
printf '\033[1;30mlooking for jinja2 in [%s]\033[0m\n' "$_py" >&2
|
||||||
$_py -c 'import jinja2' 2>/dev/null || continue
|
$_py -c 'import jinja2' 2>/dev/null || continue
|
||||||
printf '%s\n' "$_py"
|
printf '%s\n' "$_py"
|
||||||
mv $dir/{,x.}jinja2
|
mv $dir/{,x.}dep-j2
|
||||||
break
|
break
|
||||||
done)"
|
done)"
|
||||||
|
|
||||||
|
|||||||
209
scripts/test/smoketest.py
Normal file
209
scripts/test/smoketest.py
Normal file
@@ -0,0 +1,209 @@
|
|||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import time
|
||||||
|
import shlex
|
||||||
|
import shutil
|
||||||
|
import signal
|
||||||
|
import tempfile
|
||||||
|
import requests
|
||||||
|
import threading
|
||||||
|
import subprocess as sp
|
||||||
|
|
||||||
|
|
||||||
|
CPP = []
|
||||||
|
|
||||||
|
|
||||||
|
class Cpp(object):
|
||||||
|
def __init__(self, args):
|
||||||
|
args = [sys.executable, "-m", "copyparty"] + args
|
||||||
|
print(" ".join([shlex.quote(x) for x in args]))
|
||||||
|
|
||||||
|
self.ls_pre = set(list(os.listdir()))
|
||||||
|
self.p = sp.Popen(args)
|
||||||
|
# , stdout=sp.PIPE, stderr=sp.PIPE)
|
||||||
|
|
||||||
|
self.t = threading.Thread(target=self._run)
|
||||||
|
self.t.daemon = True
|
||||||
|
self.t.start()
|
||||||
|
|
||||||
|
def _run(self):
|
||||||
|
self.so, self.se = self.p.communicate()
|
||||||
|
|
||||||
|
def stop(self, wait):
|
||||||
|
if wait:
|
||||||
|
os.kill(self.p.pid, signal.SIGINT)
|
||||||
|
self.t.join(timeout=2)
|
||||||
|
else:
|
||||||
|
self.p.kill() # macos py3.8
|
||||||
|
|
||||||
|
def clean(self):
|
||||||
|
t = os.listdir()
|
||||||
|
for f in t:
|
||||||
|
if f not in self.ls_pre and f.startswith("up."):
|
||||||
|
os.unlink(f)
|
||||||
|
|
||||||
|
def await_idle(self, ub, timeout):
|
||||||
|
req = ["scanning</td><td>False", "hash-q</td><td>0", "tag-q</td><td>0"]
|
||||||
|
lim = int(timeout * 10)
|
||||||
|
u = ub + "?h"
|
||||||
|
for n in range(lim):
|
||||||
|
try:
|
||||||
|
time.sleep(0.1)
|
||||||
|
r = requests.get(u, timeout=0.1)
|
||||||
|
for x in req:
|
||||||
|
if x not in r.text:
|
||||||
|
print("ST: {}/{} miss {}".format(n, lim, x))
|
||||||
|
raise Exception()
|
||||||
|
print("ST: idle")
|
||||||
|
return
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def tc1():
|
||||||
|
ub = "http://127.0.0.1:4321/"
|
||||||
|
td = os.path.join("srv", "smoketest")
|
||||||
|
try:
|
||||||
|
shutil.rmtree(td)
|
||||||
|
except:
|
||||||
|
if os.path.exists(td):
|
||||||
|
raise
|
||||||
|
|
||||||
|
for _ in range(10):
|
||||||
|
try:
|
||||||
|
os.mkdir(td)
|
||||||
|
except:
|
||||||
|
time.sleep(0.1) # win10
|
||||||
|
|
||||||
|
assert os.path.exists(td)
|
||||||
|
|
||||||
|
vidp = os.path.join(tempfile.gettempdir(), "smoketest.h264")
|
||||||
|
if not os.path.exists(vidp):
|
||||||
|
cmd = "ffmpeg -f lavfi -i testsrc=48x32:3 -t 1 -c:v libx264 -tune animation -preset veryslow -crf 69"
|
||||||
|
sp.check_call(cmd.split(" ") + [vidp])
|
||||||
|
|
||||||
|
with open(vidp, "rb") as f:
|
||||||
|
ovid = f.read()
|
||||||
|
|
||||||
|
args = [
|
||||||
|
"-p4321",
|
||||||
|
"-e2dsa",
|
||||||
|
"-e2tsr",
|
||||||
|
"--no-mutagen",
|
||||||
|
"--th-ff-jpg",
|
||||||
|
"--hist",
|
||||||
|
os.path.join(td, "dbm"),
|
||||||
|
]
|
||||||
|
pdirs = []
|
||||||
|
hpaths = {}
|
||||||
|
|
||||||
|
for d1 in ["r", "w", "a"]:
|
||||||
|
pdirs.append("{}/{}".format(td, d1))
|
||||||
|
pdirs.append("{}/{}/j".format(td, d1))
|
||||||
|
for d2 in ["r", "w", "a"]:
|
||||||
|
d = os.path.join(td, d1, "j", d2)
|
||||||
|
pdirs.append(d)
|
||||||
|
os.makedirs(d)
|
||||||
|
|
||||||
|
pdirs = [x.replace("\\", "/") for x in pdirs]
|
||||||
|
udirs = [x.split("/", 2)[2] for x in pdirs]
|
||||||
|
perms = [x.rstrip("j/")[-1] for x in pdirs]
|
||||||
|
for pd, ud, p in zip(pdirs, udirs, perms):
|
||||||
|
if ud[-1] == "j":
|
||||||
|
continue
|
||||||
|
|
||||||
|
hp = None
|
||||||
|
if pd.endswith("st/a"):
|
||||||
|
hp = hpaths[ud] = os.path.join(td, "db1")
|
||||||
|
elif pd[:-1].endswith("a/j/"):
|
||||||
|
hpaths[ud] = os.path.join(td, "dbm")
|
||||||
|
hp = None
|
||||||
|
else:
|
||||||
|
hp = "-"
|
||||||
|
hpaths[ud] = os.path.join(pd, ".hist")
|
||||||
|
|
||||||
|
arg = "{}:{}:{}".format(pd, ud, p, hp)
|
||||||
|
if hp:
|
||||||
|
arg += ":chist=" + hp
|
||||||
|
|
||||||
|
args += ["-v", arg]
|
||||||
|
|
||||||
|
# return
|
||||||
|
cpp = Cpp(args)
|
||||||
|
CPP.append(cpp)
|
||||||
|
cpp.await_idle(ub, 3)
|
||||||
|
|
||||||
|
for d in udirs:
|
||||||
|
vid = ovid + "\n{}".format(d).encode("utf-8")
|
||||||
|
try:
|
||||||
|
requests.post(ub + d, data={"act": "bput"}, files={"f": ("a.h264", vid)})
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
cpp.clean()
|
||||||
|
|
||||||
|
# GET permission
|
||||||
|
for d, p in zip(udirs, perms):
|
||||||
|
u = "{}{}/a.h264".format(ub, d)
|
||||||
|
r = requests.get(u)
|
||||||
|
ok = bool(r)
|
||||||
|
if ok != (p in ["a"]):
|
||||||
|
raise Exception("get {} with perm {} at {}".format(ok, p, u))
|
||||||
|
|
||||||
|
# stat filesystem
|
||||||
|
for d, p in zip(pdirs, perms):
|
||||||
|
u = "{}/a.h264".format(d)
|
||||||
|
ok = os.path.exists(u)
|
||||||
|
if ok != (p in ["a", "w"]):
|
||||||
|
raise Exception("stat {} with perm {} at {}".format(ok, p, u))
|
||||||
|
|
||||||
|
# GET thumbnail, vreify contents
|
||||||
|
for d, p in zip(udirs, perms):
|
||||||
|
u = "{}{}/a.h264?th=j".format(ub, d)
|
||||||
|
r = requests.get(u)
|
||||||
|
ok = bool(r and r.content[:3] == b"\xff\xd8\xff")
|
||||||
|
if ok != (p in ["a"]):
|
||||||
|
raise Exception("thumb {} with perm {} at {}".format(ok, p, u))
|
||||||
|
|
||||||
|
# check tags
|
||||||
|
cpp.await_idle(ub, 5)
|
||||||
|
for d, p in zip(udirs, perms):
|
||||||
|
u = "{}{}?ls".format(ub, d)
|
||||||
|
r = requests.get(u)
|
||||||
|
j = r.json() if r else False
|
||||||
|
tag = None
|
||||||
|
if j:
|
||||||
|
for f in j["files"]:
|
||||||
|
tag = tag or f["tags"].get("res")
|
||||||
|
|
||||||
|
r_ok = bool(j)
|
||||||
|
w_ok = bool(r_ok and j.get("files"))
|
||||||
|
|
||||||
|
if not r_ok or w_ok != (p in ["a"]):
|
||||||
|
raise Exception("ls {} with perm {} at {}".format(ok, p, u))
|
||||||
|
|
||||||
|
if (tag and p != "a") or (not tag and p == "a"):
|
||||||
|
raise Exception("tag {} with perm {} at {}".format(tag, p, u))
|
||||||
|
|
||||||
|
if tag is not None and tag != "48x32":
|
||||||
|
raise Exception("tag [{}] at {}".format(tag, u))
|
||||||
|
|
||||||
|
cpp.stop(True)
|
||||||
|
|
||||||
|
|
||||||
|
def run(tc):
|
||||||
|
try:
|
||||||
|
tc()
|
||||||
|
finally:
|
||||||
|
try:
|
||||||
|
CPP[0].stop(False)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
run(tc1)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
53
setup.py
53
setup.py
@@ -5,23 +5,8 @@ from __future__ import print_function
|
|||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
from shutil import rmtree
|
from shutil import rmtree
|
||||||
|
|
||||||
setuptools_available = True
|
|
||||||
try:
|
|
||||||
# need setuptools to build wheel
|
|
||||||
from setuptools import setup, Command, find_packages
|
from setuptools import setup, Command, find_packages
|
||||||
|
|
||||||
except ImportError:
|
|
||||||
# works in a pinch
|
|
||||||
setuptools_available = False
|
|
||||||
from distutils.core import setup, Command
|
|
||||||
|
|
||||||
from distutils.spawn import spawn
|
|
||||||
|
|
||||||
if "bdist_wheel" in sys.argv and not setuptools_available:
|
|
||||||
print("cannot build wheel without setuptools")
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
|
|
||||||
NAME = "copyparty"
|
NAME = "copyparty"
|
||||||
VERSION = None
|
VERSION = None
|
||||||
@@ -100,9 +85,8 @@ args = {
|
|||||||
"author_email": "copyparty@ocv.me",
|
"author_email": "copyparty@ocv.me",
|
||||||
"url": "https://github.com/9001/copyparty",
|
"url": "https://github.com/9001/copyparty",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"data_files": data_files,
|
|
||||||
"classifiers": [
|
"classifiers": [
|
||||||
"Development Status :: 3 - Alpha",
|
"Development Status :: 4 - Beta",
|
||||||
"License :: OSI Approved :: MIT License",
|
"License :: OSI Approved :: MIT License",
|
||||||
"Programming Language :: Python",
|
"Programming Language :: Python",
|
||||||
"Programming Language :: Python :: 2",
|
"Programming Language :: Python :: 2",
|
||||||
@@ -120,35 +104,16 @@ args = {
|
|||||||
"Environment :: Console",
|
"Environment :: Console",
|
||||||
"Environment :: No Input/Output (Daemon)",
|
"Environment :: No Input/Output (Daemon)",
|
||||||
"Topic :: Communications :: File Sharing",
|
"Topic :: Communications :: File Sharing",
|
||||||
|
"Topic :: Internet :: WWW/HTTP :: HTTP Servers",
|
||||||
],
|
],
|
||||||
|
"include_package_data": True,
|
||||||
|
"data_files": data_files,
|
||||||
|
"packages": find_packages(),
|
||||||
|
"install_requires": ["jinja2"],
|
||||||
|
"extras_require": {"thumbnails": ["Pillow"], "audiotags": ["mutagen"]},
|
||||||
|
"entry_points": {"console_scripts": ["copyparty = copyparty.__main__:main"]},
|
||||||
|
"scripts": ["bin/copyparty-fuse.py"],
|
||||||
"cmdclass": {"clean2": clean2},
|
"cmdclass": {"clean2": clean2},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
if setuptools_available:
|
|
||||||
args.update(
|
|
||||||
{
|
|
||||||
"packages": find_packages(),
|
|
||||||
"install_requires": ["jinja2"],
|
|
||||||
"extras_require": {"thumbnails": ["Pillow"]},
|
|
||||||
"include_package_data": True,
|
|
||||||
"entry_points": {
|
|
||||||
"console_scripts": ["copyparty = copyparty.__main__:main"]
|
|
||||||
},
|
|
||||||
"scripts": ["bin/copyparty-fuse.py"],
|
|
||||||
}
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
args.update(
|
|
||||||
{
|
|
||||||
"packages": ["copyparty", "copyparty.stolen"],
|
|
||||||
"scripts": ["bin/copyparty-fuse.py"],
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
# import pprint
|
|
||||||
# pprint.PrettyPrinter().pprint(args)
|
|
||||||
# sys.exit(0)
|
|
||||||
|
|
||||||
setup(**args)
|
setup(**args)
|
||||||
|
|||||||
@@ -8,13 +8,13 @@ import time
|
|||||||
import shutil
|
import shutil
|
||||||
import pprint
|
import pprint
|
||||||
import tarfile
|
import tarfile
|
||||||
|
import tempfile
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from argparse import Namespace
|
from argparse import Namespace
|
||||||
from copyparty.authsrv import AuthSrv
|
|
||||||
from copyparty.httpcli import HttpCli
|
|
||||||
|
|
||||||
from tests import util as tu
|
from tests import util as tu
|
||||||
|
from copyparty.authsrv import AuthSrv
|
||||||
|
from copyparty.httpcli import HttpCli
|
||||||
|
|
||||||
|
|
||||||
def hdr(query):
|
def hdr(query):
|
||||||
@@ -32,21 +32,27 @@ class Cfg(Namespace):
|
|||||||
no_zip=False,
|
no_zip=False,
|
||||||
no_scandir=False,
|
no_scandir=False,
|
||||||
no_sendfile=True,
|
no_sendfile=True,
|
||||||
|
no_rescan=True,
|
||||||
|
ihead=False,
|
||||||
nih=True,
|
nih=True,
|
||||||
mtp=[],
|
mtp=[],
|
||||||
mte="a",
|
mte="a",
|
||||||
|
hist=None,
|
||||||
|
no_hash=False,
|
||||||
**{k: False for k in "e2d e2ds e2dsa e2t e2ts e2tsr".split()}
|
**{k: False for k in "e2d e2ds e2dsa e2t e2ts e2tsr".split()}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class TestHttpCli(unittest.TestCase):
|
class TestHttpCli(unittest.TestCase):
|
||||||
def test(self):
|
def setUp(self):
|
||||||
td = os.path.join(tu.get_ramdisk(), "vfs")
|
self.td = tu.get_ramdisk()
|
||||||
try:
|
|
||||||
shutil.rmtree(td)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
def tearDown(self):
|
||||||
|
os.chdir(tempfile.gettempdir())
|
||||||
|
shutil.rmtree(self.td)
|
||||||
|
|
||||||
|
def test(self):
|
||||||
|
td = os.path.join(self.td, "vfs")
|
||||||
os.mkdir(td)
|
os.mkdir(td)
|
||||||
os.chdir(td)
|
os.chdir(td)
|
||||||
|
|
||||||
@@ -95,7 +101,7 @@ class TestHttpCli(unittest.TestCase):
|
|||||||
pprint.pprint(vcfg)
|
pprint.pprint(vcfg)
|
||||||
|
|
||||||
self.args = Cfg(v=vcfg, a=["o:o", "x:x"])
|
self.args = Cfg(v=vcfg, a=["o:o", "x:x"])
|
||||||
self.auth = AuthSrv(self.args, self.log)
|
self.asrv = AuthSrv(self.args, self.log)
|
||||||
vfiles = [x for x in allfiles if x.startswith(top)]
|
vfiles = [x for x in allfiles if x.startswith(top)]
|
||||||
for fp in vfiles:
|
for fp in vfiles:
|
||||||
rok, wok = self.can_rw(fp)
|
rok, wok = self.can_rw(fp)
|
||||||
@@ -184,12 +190,12 @@ class TestHttpCli(unittest.TestCase):
|
|||||||
def put(self, url):
|
def put(self, url):
|
||||||
buf = "PUT /{0} HTTP/1.1\r\nCookie: cppwd=o\r\nConnection: close\r\nContent-Length: {1}\r\n\r\nok {0}\n"
|
buf = "PUT /{0} HTTP/1.1\r\nCookie: cppwd=o\r\nConnection: close\r\nContent-Length: {1}\r\n\r\nok {0}\n"
|
||||||
buf = buf.format(url, len(url) + 4).encode("utf-8")
|
buf = buf.format(url, len(url) + 4).encode("utf-8")
|
||||||
conn = tu.VHttpConn(self.args, self.auth, self.log, buf)
|
conn = tu.VHttpConn(self.args, self.asrv, self.log, buf)
|
||||||
HttpCli(conn).run()
|
HttpCli(conn).run()
|
||||||
return conn.s._reply.decode("utf-8").split("\r\n\r\n", 1)
|
return conn.s._reply.decode("utf-8").split("\r\n\r\n", 1)
|
||||||
|
|
||||||
def curl(self, url, binary=False):
|
def curl(self, url, binary=False):
|
||||||
conn = tu.VHttpConn(self.args, self.auth, self.log, hdr(url))
|
conn = tu.VHttpConn(self.args, self.asrv, self.log, hdr(url))
|
||||||
HttpCli(conn).run()
|
HttpCli(conn).run()
|
||||||
if binary:
|
if binary:
|
||||||
h, b = conn.s._reply.split(b"\r\n\r\n", 1)
|
h, b = conn.s._reply.split(b"\r\n\r\n", 1)
|
||||||
|
|||||||
@@ -7,24 +7,30 @@ import json
|
|||||||
import shutil
|
import shutil
|
||||||
import tempfile
|
import tempfile
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from textwrap import dedent
|
from textwrap import dedent
|
||||||
from argparse import Namespace
|
from argparse import Namespace
|
||||||
from copyparty.authsrv import AuthSrv
|
|
||||||
from copyparty import util
|
|
||||||
|
|
||||||
from tests import util as tu
|
from tests import util as tu
|
||||||
|
from copyparty.authsrv import AuthSrv, VFS
|
||||||
|
from copyparty import util
|
||||||
|
|
||||||
|
|
||||||
class Cfg(Namespace):
|
class Cfg(Namespace):
|
||||||
def __init__(self, a=[], v=[], c=None):
|
def __init__(self, a=[], v=[], c=None):
|
||||||
ex = {k: False for k in "e2d e2ds e2dsa e2t e2ts e2tsr".split()}
|
ex = {k: False for k in "e2d e2ds e2dsa e2t e2ts e2tsr".split()}
|
||||||
ex["mtp"] = []
|
ex2 = {"mtp": [], "mte": "a", "hist": None, "no_hash": False}
|
||||||
ex["mte"] = "a"
|
ex.update(ex2)
|
||||||
super(Cfg, self).__init__(a=a, v=v, c=c, **ex)
|
super(Cfg, self).__init__(a=a, v=v, c=c, **ex)
|
||||||
|
|
||||||
|
|
||||||
class TestVFS(unittest.TestCase):
|
class TestVFS(unittest.TestCase):
|
||||||
|
def setUp(self):
|
||||||
|
self.td = tu.get_ramdisk()
|
||||||
|
|
||||||
|
def tearDown(self):
|
||||||
|
os.chdir(tempfile.gettempdir())
|
||||||
|
shutil.rmtree(self.td)
|
||||||
|
|
||||||
def dump(self, vfs):
|
def dump(self, vfs):
|
||||||
print(json.dumps(vfs, indent=4, sort_keys=True, default=lambda o: o.__dict__))
|
print(json.dumps(vfs, indent=4, sort_keys=True, default=lambda o: o.__dict__))
|
||||||
|
|
||||||
@@ -41,6 +47,7 @@ class TestVFS(unittest.TestCase):
|
|||||||
self.assertEqual(util.undot(query), response)
|
self.assertEqual(util.undot(query), response)
|
||||||
|
|
||||||
def ls(self, vfs, vpath, uname):
|
def ls(self, vfs, vpath, uname):
|
||||||
|
# type: (VFS, str, str) -> tuple[str, str, str]
|
||||||
"""helper for resolving and listing a folder"""
|
"""helper for resolving and listing a folder"""
|
||||||
vn, rem = vfs.get(vpath, uname, True, False)
|
vn, rem = vfs.get(vpath, uname, True, False)
|
||||||
r1 = vn.ls(rem, uname, False)
|
r1 = vn.ls(rem, uname, False)
|
||||||
@@ -55,12 +62,7 @@ class TestVFS(unittest.TestCase):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
def test(self):
|
def test(self):
|
||||||
td = os.path.join(tu.get_ramdisk(), "vfs")
|
td = os.path.join(self.td, "vfs")
|
||||||
try:
|
|
||||||
shutil.rmtree(td)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
|
|
||||||
os.mkdir(td)
|
os.mkdir(td)
|
||||||
os.chdir(td)
|
os.chdir(td)
|
||||||
|
|
||||||
@@ -111,13 +113,13 @@ class TestVFS(unittest.TestCase):
|
|||||||
n = vfs.nodes["a"]
|
n = vfs.nodes["a"]
|
||||||
self.assertEqual(len(vfs.nodes), 1)
|
self.assertEqual(len(vfs.nodes), 1)
|
||||||
self.assertEqual(n.vpath, "a")
|
self.assertEqual(n.vpath, "a")
|
||||||
self.assertEqual(n.realpath, td + "/a")
|
self.assertEqual(n.realpath, os.path.join(td, "a"))
|
||||||
self.assertEqual(n.uread, ["*", "k"])
|
self.assertEqual(n.uread, ["*", "k"])
|
||||||
self.assertEqual(n.uwrite, ["k"])
|
self.assertEqual(n.uwrite, ["k"])
|
||||||
n = n.nodes["ac"]
|
n = n.nodes["ac"]
|
||||||
self.assertEqual(len(vfs.nodes), 1)
|
self.assertEqual(len(vfs.nodes), 1)
|
||||||
self.assertEqual(n.vpath, "a/ac")
|
self.assertEqual(n.vpath, "a/ac")
|
||||||
self.assertEqual(n.realpath, td + "/a/ac")
|
self.assertEqual(n.realpath, os.path.join(td, "a", "ac"))
|
||||||
self.assertEqual(n.uread, ["*", "k"])
|
self.assertEqual(n.uread, ["*", "k"])
|
||||||
self.assertEqual(n.uwrite, ["k"])
|
self.assertEqual(n.uwrite, ["k"])
|
||||||
n = n.nodes["acb"]
|
n = n.nodes["acb"]
|
||||||
@@ -227,7 +229,7 @@ class TestVFS(unittest.TestCase):
|
|||||||
self.assertEqual(list(v1), list(v2))
|
self.assertEqual(list(v1), list(v2))
|
||||||
|
|
||||||
# config file parser
|
# config file parser
|
||||||
cfg_path = os.path.join(tu.get_ramdisk(), "test.cfg")
|
cfg_path = os.path.join(self.td, "test.cfg")
|
||||||
with open(cfg_path, "wb") as f:
|
with open(cfg_path, "wb") as f:
|
||||||
f.write(
|
f.write(
|
||||||
dedent(
|
dedent(
|
||||||
@@ -249,7 +251,7 @@ class TestVFS(unittest.TestCase):
|
|||||||
n = au.vfs
|
n = au.vfs
|
||||||
# root was not defined, so PWD with no access to anyone
|
# root was not defined, so PWD with no access to anyone
|
||||||
self.assertEqual(n.vpath, "")
|
self.assertEqual(n.vpath, "")
|
||||||
self.assertEqual(n.realpath, td)
|
self.assertEqual(n.realpath, None)
|
||||||
self.assertEqual(n.uread, [])
|
self.assertEqual(n.uread, [])
|
||||||
self.assertEqual(n.uwrite, [])
|
self.assertEqual(n.uwrite, [])
|
||||||
self.assertEqual(len(n.nodes), 1)
|
self.assertEqual(len(n.nodes), 1)
|
||||||
@@ -260,6 +262,4 @@ class TestVFS(unittest.TestCase):
|
|||||||
self.assertEqual(n.uwrite, ["asd"])
|
self.assertEqual(n.uwrite, ["asd"])
|
||||||
self.assertEqual(len(n.nodes), 0)
|
self.assertEqual(len(n.nodes), 0)
|
||||||
|
|
||||||
os.chdir(tempfile.gettempdir())
|
|
||||||
shutil.rmtree(td)
|
|
||||||
os.unlink(cfg_path)
|
os.unlink(cfg_path)
|
||||||
|
|||||||
@@ -1,16 +1,36 @@
|
|||||||
import os
|
import os
|
||||||
|
import sys
|
||||||
import time
|
import time
|
||||||
|
import shutil
|
||||||
import jinja2
|
import jinja2
|
||||||
import tempfile
|
import tempfile
|
||||||
|
import platform
|
||||||
import subprocess as sp
|
import subprocess as sp
|
||||||
|
|
||||||
from copyparty.util import Unrecv
|
|
||||||
|
|
||||||
|
WINDOWS = platform.system() == "Windows"
|
||||||
|
ANYWIN = WINDOWS or sys.platform in ["msys"]
|
||||||
|
MACOS = platform.system() == "Darwin"
|
||||||
|
|
||||||
J2_ENV = jinja2.Environment(loader=jinja2.BaseLoader)
|
J2_ENV = jinja2.Environment(loader=jinja2.BaseLoader)
|
||||||
J2_FILES = J2_ENV.from_string("{{ files|join('\n') }}")
|
J2_FILES = J2_ENV.from_string("{{ files|join('\n') }}")
|
||||||
|
|
||||||
|
|
||||||
|
def nah(*a, **ka):
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
if MACOS:
|
||||||
|
import posixpath
|
||||||
|
|
||||||
|
posixpath.islink = nah
|
||||||
|
os.path.islink = nah
|
||||||
|
# 25% faster; until any tests do symlink stuff
|
||||||
|
|
||||||
|
|
||||||
|
from copyparty.util import Unrecv
|
||||||
|
|
||||||
|
|
||||||
def runcmd(*argv):
|
def runcmd(*argv):
|
||||||
p = sp.Popen(argv, stdout=sp.PIPE, stderr=sp.PIPE)
|
p = sp.Popen(argv, stdout=sp.PIPE, stderr=sp.PIPE)
|
||||||
stdout, stderr = p.communicate()
|
stdout, stderr = p.communicate()
|
||||||
@@ -28,18 +48,25 @@ def chkcmd(*argv):
|
|||||||
|
|
||||||
|
|
||||||
def get_ramdisk():
|
def get_ramdisk():
|
||||||
|
def subdir(top):
|
||||||
|
ret = os.path.join(top, "cptd-{}".format(os.getpid()))
|
||||||
|
shutil.rmtree(ret, True)
|
||||||
|
os.mkdir(ret)
|
||||||
|
return ret
|
||||||
|
|
||||||
for vol in ["/dev/shm", "/Volumes/cptd"]: # nosec (singleton test)
|
for vol in ["/dev/shm", "/Volumes/cptd"]: # nosec (singleton test)
|
||||||
if os.path.exists(vol):
|
if os.path.exists(vol):
|
||||||
return vol
|
return subdir(vol)
|
||||||
|
|
||||||
if os.path.exists("/Volumes"):
|
if os.path.exists("/Volumes"):
|
||||||
devname, _ = chkcmd("hdiutil", "attach", "-nomount", "ram://32768")
|
# hdiutil eject /Volumes/cptd/
|
||||||
|
devname, _ = chkcmd("hdiutil", "attach", "-nomount", "ram://131072")
|
||||||
devname = devname.strip()
|
devname = devname.strip()
|
||||||
print("devname: [{}]".format(devname))
|
print("devname: [{}]".format(devname))
|
||||||
for _ in range(10):
|
for _ in range(10):
|
||||||
try:
|
try:
|
||||||
_, _ = chkcmd("diskutil", "eraseVolume", "HFS+", "cptd", devname)
|
_, _ = chkcmd("diskutil", "eraseVolume", "HFS+", "cptd", devname)
|
||||||
return "/Volumes/cptd"
|
return subdir("/Volumes/cptd")
|
||||||
except Exception as ex:
|
except Exception as ex:
|
||||||
print(repr(ex))
|
print(repr(ex))
|
||||||
time.sleep(0.25)
|
time.sleep(0.25)
|
||||||
@@ -50,7 +77,7 @@ def get_ramdisk():
|
|||||||
try:
|
try:
|
||||||
os.mkdir(ret)
|
os.mkdir(ret)
|
||||||
finally:
|
finally:
|
||||||
return ret
|
return subdir(ret)
|
||||||
|
|
||||||
|
|
||||||
class NullBroker(object):
|
class NullBroker(object):
|
||||||
@@ -83,15 +110,19 @@ class VHttpSrv(object):
|
|||||||
|
|
||||||
|
|
||||||
class VHttpConn(object):
|
class VHttpConn(object):
|
||||||
def __init__(self, args, auth, log, buf):
|
def __init__(self, args, asrv, log, buf):
|
||||||
self.s = VSock(buf)
|
self.s = VSock(buf)
|
||||||
self.sr = Unrecv(self.s)
|
self.sr = Unrecv(self.s)
|
||||||
self.addr = ("127.0.0.1", "42069")
|
self.addr = ("127.0.0.1", "42069")
|
||||||
self.args = args
|
self.args = args
|
||||||
self.auth = auth
|
self.asrv = asrv
|
||||||
|
self.is_mp = False
|
||||||
self.log_func = log
|
self.log_func = log
|
||||||
self.log_src = "a"
|
self.log_src = "a"
|
||||||
|
self.lf_url = None
|
||||||
self.hsrv = VHttpSrv()
|
self.hsrv = VHttpSrv()
|
||||||
self.nbyte = 0
|
self.nbyte = 0
|
||||||
self.workload = 0
|
self.workload = 0
|
||||||
|
self.ico = None
|
||||||
|
self.thumbcli = None
|
||||||
self.t0 = time.time()
|
self.t0 = time.time()
|
||||||
Reference in New Issue
Block a user