RogueRose
International Hazard
Posts: 1590
Registered: 16-6-2014
Member Is Offline
|
|
Anyone scrape/save Youtube video comments? I need a good method
I download YT vids that are good and often the comment sections are full of very useful information but I haven't found a good program that will save
the comments, and IDK what format would be best. I've seen some that save in JSON or CSV and I guess these can be imported into other programs but it
is still hard to view them.
Anyone come across this issue and found a good solution?
|
|
mayko
International Hazard
Posts: 1218
Registered: 17-1-2013
Location: Carrboro, NC
Member Is Offline
Mood: anomalous (Euclid class)
|
|
What does ease of reading mean here? Most spreadsheet software will open a CSV without problem; I just did it in LibreOffice Calc. They can also be
easily processed on the command line, including using tools like tr to convert commas to whitespace for less cluttered reading.
al-khemie is not a terrorist organization
"Chemicals, chemicals... I need chemicals!" - George Hayduke
"Wubbalubba dub-dub!" - Rick Sanchez
|
|
j_sum1
Administrator
Posts: 6278
Registered: 4-10-2014
Location: Unmoved
Member Is Offline
Mood: Organised
|
|
Chemplayer said he archived his but I don't know how he did it.
|
|
JJay
International Hazard
Posts: 3440
Registered: 15-10-2015
Member Is Offline
|
|
YouTube has an API, and it allows retrieving comments: https://developers.google.com/youtube/v3/guides/implementati...
It would be easier and more elegant to use the API than to scrape. Practically every programming language can parse JSON.
|
|
Tsjerk
International Hazard
Posts: 3028
Registered: 20-4-2005
Location: Netherlands
Member Is Offline
Mood: Mood
|
|
I would go for the API, with for example Postman you have a nice GUI for calling API's, and the JSON can be made readable with code editors like e.g.
Brackets (with a beautify plugin).
|
|
DrEvidence
Harmless
Posts: 37
Registered: 25-4-2018
Member Is Offline
Mood: No Mood
|
|
I sometimes collect some dynamic stuff like news, comments, stats...in fear that somebody would die or dissappear or ban or stop providing
content/electricity/internet...for example if somebody becomes sick or angry or tired or sad or arrested or no longer interested. There are many
youtube channels which got banned and many webpages dissappeared or siezed by the fbi.
I don't know what's wrong with doing that all manually. I saved...let me count...6683 pages as mhtml files manually, from 432 domains (websites). Plus
csv is so standard and small size and fast. I use nirsoft csvfileview as viewer, and make most csv manually using notepad++, simply copy full first
line and modify each word with your content in new line...in most cases need to put "" for example for youtube comments or large texts.
Microsoft's and various offices are too slow to open csv file, and are worth to use only if you use less larger file, not vice versa.
Hell, if you don't have time to do it all manually, then you it means don't have time to read it all later.
Hell, I don't know what API is, sound like something complicated for programmers. I will only become programmer if i simplify those complications.
|
|