Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> are so evenly matched

It's because the real value of the models is in what we (humanity) fed them, and all of them have eaten the same thing for free.



That's why the frontier LLM companies are now spending a lot more to license exclusive proprietary training data from private sources in order to gain a quality edge in certain business domains.


But those holding said proprietary data have figured out they’re holding the cards now and have gotten a lot smarter recently. Companies are being very careful about what gets used for inference vs what they allow to be used for training.

I don’t see the core models getting dramatically better from where they are now. We’ve clearly hit a plateau.


Really? I mean I see regularly as I'm coding how much better it could be simply by running obvious prompts for me.

When I use the planning mode and then code the success rate is much higher. When I ask it to work on specific isolated chunks of code with clear success/failure modes the success rate is again much higher.

Now imagine a world where it recognizes that from my simple throw away non specific prompt. If it was able to fire off 20 different prompts in quick succession it could easily cut my time spent in front of the screen by a third.

The patterns are obvious but they don't do that right now because it's a lot of compute.

We'll be looking at this time where there's a progress bar showing context space the way we look at the Turbo button.

Because the truth is to get the baseline I'm talking about is a finite amount of compute at a certain point.


so can it be the one that gets ahead on having people go find things for them - https://news.ycombinator.com/item?id=47285283


Interesting


That sounds like spin to me. If there were a clear "quality edge" in "certain business domains" stemming from "exclusive proprietary data", someone would have been exploiting it already using meat computers.

But no, businesses are dumb. They always have been. Existing businesses get disrupted by new ideas and new technology all the time. This very site is a temple to disruption!

Proprietary advantage is, 99.999% of the time, just structural advantage. You can't compete with Procter & Gamble because they already built their brands and factories and supply chains and you'd have to do all that from scratch while selling cheaper products as upstart value options. And there's not enough money in consumer junk to make that worth it.

But if you did have funding and wanted to beat them on first principles? Would you really start by training an LLM on what they're already doing? No, you'd throw money at a bunch of hackers from YC. Duh.


Frontier labs are paying the same constellation of firms offering proprietary data and access to experts in their fields to train LLMs.

They are neck-and-neck only because they are participating in the arms race. The only other way to keep up is mass-distillation, which could prove to be fragile (so far it seems to be sustainable).


Meh. I think there's basically no benefit shown so far to careful curation. That's where we've been in machine learning for three decades, after all. Also recognize that the Great Leap Forward of LLMs was when they got big enough to abandon that strategy and just slurp in the Library of All The Junk.

I think one needs to at least recognize the possibility that... there just isn't any more data for training. We've done it all. The models we have today have already distilled all of the output of human cleverness throughout history. If there's more data to be had, we need to make it the hard way.


Ok, maybe pretraining is now complete and solved. Next up: post-training, reinforcement learning, engineering RL environments for realistic problem solving, recording data online during use, then offline simulation of how it could have gone better and faster, distilling that into the next model etc. etc. There's still decades worth of progress to be made this way.


" There's still decades worth of progress to be made this way."

That's not true. Moreover the progress can slow to a crawl where it's barely noticeable. And in that world the humans continues to stay ahead - that's the magic of humans. To be aware of surroundings and adapt sufficiently whilst taking advantage of tools and leveraging them.


This is an interesting theoretical statement that does not survive a collision with reality. The long-tail expert RHLF training is effective. We have seen significant employment impact to call center employees. This does not mean its progress will be cheap or immediate.


I think this is where we are at, too.

But if you say stuff like this on here you get down voted. Why?


The quality edge hasn't shown up yet. If this strategy actually works then the quality improvements will only become apparent in the next round of major LLM updates. There's a lot of valuable training data locked up behind corporate firewalls. But this is all somewhat speculative for now.


To stop this, I today put most of my Amazon Redshift research web-site behind a basic auth username/password wall.

It's all remains free, but you need to email me for a username and password.

If I put in time and effort to make content and OpenAI et al copy it and sell it through their LLM such that no one comes to me any more, then plainly it makes no sense for me to create that content; and then it would not exist for OpenAI to take, or for anyone else. We all lose.

It seems parasitic.


An AI is more likely than me to take the time to send you an email for requesting access - I'm too lazy.


I think a better approach would be to have a login form and just say "the password is 1234" or whatever.

Virtually no scraper has logic to handle that sort of situation, but it's trivial for humans. Way easier than an LLM


Not true, even Windows Defender is capable of extracting "the password is 1234" from context like emails or webpages.


Please add Internet Archive's bot to your auto-allows, at least. Their bot is presumably well behaved, and for public benefit.


I'm about to ask IA to remove my content!

The reason is that I expect LLM bots to be crawling IA.


To be more precise, they all stole the same stuff. I have no empathy for these crooks.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: