Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I don't see how you can stop the LLMs ingesting any poison either, because they're filling up the internet with low-value crap as fast as they possibly can. All that junk is poisonous to training new models. The wellspring of value once provided by sites like StackoverFlow is now all but dried up. AI culture is devaluing at an incredible rate as it churns out copied and copies and copies and more copies of the same worthless junk.




The big labs spend a ton of effort on dataset curation, precisely to prevent them from ingesting poison as you put it.

It goes further than that—they do lots of testing on the dataset to find the incremental data that produces best improvements on model performance, and even train proxy models that predict whether data will improve performance or not.

“Data Quality” is usually a huge division with a big budget.


Jeez, why can't I have a data quality team filtering out AI slop!

You can... you just need to make about $100,000,000,000 USD in profits each year, that's all.

Just look at the domains. Obviously social media will get harder to do this with, maybe that's okay though. I think a simple criterion can be used: could the pre-trained LLM have come up with this itself? If so it probably doesn't have training value.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: