Ranter
Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Comments
-
Holy fuck. I wouldn't expect that all movie subtitles in the world would be 130gb
-
@retoor it's kind of impressive isn't it?
I imagine training on that data would get funky with all the time codes in it. -
@iSwimInTheC hmm, you should create a c program or so to strip all times. Python could be even too slow for that. And how much formats will be there?
It's interesting project. -
@retoor i wouldn't even believe it to be 130GB uncompressed. but as a zip? absolutely not.
-
@retoor nah, he probably just downloaded 130gb of garbage.
but let's wait until he _validated_ that all of the data he got actually consists of meaningful subtitles ;) -
vane110491y@tosensei @retoor you can download torrent from this link
https://reddit.com/r/DataHoarder/...
123gb torrent + 7gb zip from archive.org
it's single sqlite database with zip blobs in it, also names are in nice format you get some.title.(year).language.(id in database).zip, walk trough all of it takes about 2.5h in python on my NAS where you traverse each record in sqlite open zip in memory and extract subtitle skipping everything else, not much time for 130gb. -
Now I just wonder the most common phrases in media by category
If you speak to this AI would it feel like the average Disney fan?
That would be hilarious.
Related Rants
Downloaded 130gb of movie subtitles zip files.
If I find some power deep in my heart I would normalize data and launch training on generative transformer to see if it produces decent dialogues.
It will probably stop on planning phase because I’m diving deeper towards depression.
rant
dataset