>>12334
nice! you have a lot more storage than me :)
>>12313
i have a script for you. i came up with a simple way to scrape video titles from ZooX18. afaik their server doesn't provide an endpoint mapping video IDs -> titles that isn't protected by Cloudflare. however, their unprotected video search results pages (sorted by most recent) seem to list out every video on the site, so you can mass download those and scrape the HTML.
here's the URL template (* starts at 1):
https://www.zoox18.com/videos?o=mr&page=*
you'd need to mass download these pages using whatever tool you want. I found this FOSS "uGet" download manager that provides a good amount of flexibility while still being easy to use. to set it up for downloading these pages, open it, then click Category -> Properties. and bump "active downloads" up higher than the default. in the defaults for new downloads, set the Referrer to "
https://www.zoox18.com" and copy your browser's user agent into the User Agent box. finally click FIle -> Batch Downloads -> URL Sequence and set it up correctly. you can find the current max page number by opening page 1 in your browser and scrolling to the bottom (3618 when I last checked). as you download the pages, new videos may arrive, which may cause duplicates to appear, but that's okay.
i attached a "scrape_video_titles.sh" script to this post (remove the .txt from filename). it takes a directory containing downloaded HTML files and outputs the video IDs and their corresponding titles. if you run it without any arguments it prints the usage info. once you have the "id_title_map.txt", you can use something like this to bulk rename your video files:
id_title_map="id_title_map.txt" # replace with path to id_title_map.txt
videos="./videos" # replace with path to video files
for file in "$videos"/*.mp4; do
bn=$(basename "$file")
v_id=$(echo "$bn" | cut -d'_' -f1)
# "head -n1" used in case there are duplicates
v_title=$(grep "$v_id" "$id_title_map" | head -n1 | cut -d'|' -f2)
mv "$file" "$(dirname "$file")/$v_id - $v_title.mp4"
done
you could get more efficient with sorting ("sort -n") and duplicate removal ("uniq") and whatnot but it may not be necessary. i would try and mass download myself but my internet connection is dogshit. hope it works for you. thanks for your archiving work.