logoalt Hacker News

nostrademonsyesterday at 10:40 PM2 repliesview on HN

That's sorta what MetaBrainz did - they offer their whole DB as a single tarball dump, much like what Wikipedia does. I downloaded it in the order of an hour; if I need a MusicBrainz lookup, I just do a local query.

For this strategy to work, people need to actually use the DB dumps instead of just defaulting to scraping. Unfortunately scraping is trivially easy, particularly now that AI code assistants can write a working scraper in ~5-10 minutes.


Replies

8notetoday at 2:55 AM

the obvious thing would be to take down their website and only have the DB dump.

if thats the useful thing, it doesnt need the wrapper

tonyhart7today at 1:17 AM

I mean this AI data scrapper would need to scan and fetch billions of website

why would they even care over 1 single website ??? You expect instiution to care out of billions website they must scrape daily

show 1 reply