Hacker Newsnew | past | comments | ask | show | jobs | submit | bitwize's commentslogin

This is coming. In particular, without a Secure-Boot-enforced allowlist of operating systems, it will be near impossible to verify that an OS connecting to the internet complies with your locality's age verification laws, so it will soon be illegal to run a computer that does not make Secure Boot mandatory and connect it to the network.

If you're starting to think "huh, maybe that's why these age verification laws suddenly became all the rage", you're onto something. Whatever the case, "general purpose computing" is definitely cooked.


The laws in my locality place requirements on the service provider (e.g. the adult website operator), not on random computer owners or manufacturers or software vendors.

General purpose computing as it was done in the 1900s is cooked for the average user because there is no market incentive for it to exist. The actual market incentive revolves around apps as they provide user value along with the ability to deploy custom apps.

OpenAI did not make the strong specific claims about GPT2's abilities that Anthropic is making about Claude Mythos.

We're gonna find that Claude Mythos can do something like this in 255 bytes

Optimizing away one byte of this, given the source code? Yeah, could happen. Making a good 256 byte demo from scratch? No way.

I mean, give it a try?

the source is right there ;)


I mean now that the (human written) source for this is out in the wild, of course it can ;)

The new Power Mac® G4 with Velocity Engine®. So powerful, the government classifies it as a supercomputer and a potential weapon.


This leads to a well-documented phenomenon known as model collapse. You know how if you blur and sharpen an image repeatedly you eventually end up with just a rectangle of creepy, wormy spaghetti lines? You lose information on each blur, and then ask it to reconstitute the image with less information on each sharpen, until there's nothing recognizable left.

Training a model is like the blur and generating from that model is like the sharpen. Repeat enough times and enough information is lost that you're just left with "wormy spaghetti lines"—in an LLM's case, meaningless gibberish that actually pretty closely resembles the glitchy stuff said by the cores that fall off GLaDOS in Portal. I dunno, you read the paper and be the judge:

https://www.nature.com/articles/s41586-024-07566-y

To jump to the last output sample, C-f Gen 9

Of course you may be talking about the human aspect of this. Gods willing, we'll realize that our LLMs are spewing gibberish and think twice about putting them in all the things, all the time. But the scenario I fear isn't Idiocracy—it's worse: a community of humans who treat the gibberish as sacred writ, Zardoz style.


My brain keeps wanting to pronounce it Tomb Raider style, like /ˈskiː ɒn/.

Population decline? Less emissions? Haven't we reached consensus that those would be welcome today? Is it time for a pro-dark-age movement?

The world is projected to hit population decline already sometime between 2060 and 2080, so I guess the younger ones of us will find out definitively whether it's a good or bad thing.

Was just gonna say this is a great accessory to put your computer on while playing QBJ3!

Value to customer. Literally the only thing that matters.

Value isn't a one-shot, though. Value sustained over time is what matters.

Well, if unmaintainable code gets in the way of the "sustained over time" part, then that is still a real problem.


Businesses no longer seem to care about "value sustained over time"

They only seem to operate as "extract as much value as possible in a short amount of time and exit with your bag", these days


Bob+agents is going to be able to solve much more complex problems than Bob without agents.

That's the true AI revolution: not the things it can accelerate, the things it can put in reach that you wouldn't countenance doing before.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: