Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Anything articulate (hence possibly convincing) which could be «merely [guessing]» should either be locked out of consequential questions, or fixed.


We're still on that's just how it works. The LLM isn't aware of any consequence, etc. All it does is complete patterns as trained. And the data contains many instances of articulate question answering.

It is for those using the LLM to be aware of its capabilities, or not - be allowed to - use it. Like a child unaware that running their finger on a sharp knife blade will lead to a bad slice; you don't dull the blade to keep the child safe, but keep the child from the knife until they can understand and respect its capabilities.


If your prototype of the «knife» is all blade and no handle, fix it and implement the handle.

If the creation is planned, you will have also thought of the handle; if it is a serendipity, you will have to plan the handle afterwards.


Pretty sure it doesn't matter to the child whether the knife has a handle or not. They'll eventually find a way to cut themself.


It matters to the adult - who is also an user.

LLMs do not deliver (they miss important qualities related to intelligence); they are here now; so they must be superseded.

There is no excuse: they must be fixed urgently.


LLMs deliver pretty well on their intended functionality: they predict next tokens given a token history and patterns in their training data. If you want to describe that as fully intelligent, that's your call, but I personally wouldn't. And adding functionality that isn't directly related to improving token prediction is just bad practice in an already very complex creation. LLM tools exist for that reason: they're the handles, sheaths, sharpeners, etc for the knife. Teach those adults who're getting themselves cut to hold the knife by the handle and use the other accessories that improve user experience.


> given a token history and patterns in their training data. If you want to describe that as fully intelligent

No, I would call (an easy interpretation of) that an implementation of unintelligence. Following patterns is what an hearsay machine does.

The architecture you describe at the "token prediction" level collides with an architecture in which ideas get related with better justifications than frequent co-occurrance. Given that the outputs will be similar in form, and that "dubious guessers" are now in place, we are now bound to hurry towards the "certified guessers".


> Following patterns is what an hearsay machine does.

That's also how the brain works, at least partially. Primary differences are it takes and processes (trains itself on) raw sensory data instead of character tokens, and it continually does so for every conscious moment from at least birth until death.


> how the brain works, at least partially

With the difference, which have us go back to the original point, that the human mind has a crucial property of going beyond "pattern-based" intuition and check mental items lucidly and consciously.

> and it continually does so

It also persistently evaluates consciously and "store" and "learn" (which must be noted because it is the second main thing that LLMs don't do, after the problem of going past intuition).


> check mental items lucidly and consciously

Capabilities that evolved over millennia. We don't even have a decent, universally-agreed upon definition for consciousness yet.

> "store" and "learn"

Actually there are tools for that. Again, the core LLM functionality is best left on its own, and augmented on the fly with various tools which can be easily specialized and upgraded independently of the model. Consider too that the brain itself has multiple sections dedicated to different kinds of processing, instead of anything just happening anywhere.


> Capabilities that evolved over millennia

Means nothing. Now they are urgent.

> consciousness

I meant "conscious" as the wake opposed to deliriousness, as the ability "to be sure about what you have in front of you with the degree of accomplished clarity and substantially unproblematic boundaries of definition",

not as that quality that intrigues (and obsesses) less pragmatic intellectuals in directions e.g. at the "Closer to Truth" channel.

When I ask somebody, it has to be sure to a high degree. When implementing a mind, the property of "lucid conscious check" is fundamental.

> tools for that

The "consciously check, then store and learn" is structural in the proper human mental process - a high level functioning, not just a module; i.e. "it's what we do".

Which means, the basic LLM architecture is missing important features that we need if we want to implement a developed interlocutor. And we need and want that.


> Now they are urgent.

Something being urgent doesn't mean there's a known viable pathway to an ideal implementation. The brain has billions of neurons working in tandem, a scale we're nowhere near to replicating, last I checked. And there are signs pointing to scale of neural interactions as a key factor in intelligent capabilities.

> When I ask somebody, it has to be sure to a high degree.

You're looking in the wrong place if you need surety. LLMs aren't "sure" about anything, and will never be. We going in circles at this point, but if you need certitude in anything, add tools to the LLM to increase surety. For some reason you seem to be pushing for a less optimal solution to a problem that already has a decent one.

BTW seems you may be - unconsciously? - crossing a person with an LLM there.

> structural in the proper human mental process

The brain is very modularized. It has sections/lobes which specialize in core life support functions, seeing, hearing, movement, reasoning, memory, etc. That's why brain surgeons can reliably know what capabilities may be affected by their actions in a given part of the brain. And all those functions are tools in some way to something else, for example reasoning would be pretty limited without memory.

And even at a larger scope, as humans we still use tools to achieve greater surety from our mental processes. That's why we have calculators, watches, cameras, etc. And why some would type "blueberry" into a tool with proven spell checking or autocorrect capabilities and eyeball the letters for a couple seconds to confirm the number of "b"s in it. The brain as a whole is still pretty fallible with all its capacity and capabilities.


> doesn't mean there's a known viable pathway [...] solution [...] BTW seems

It means it must be researched into with high commitment. // LLMs are an emergency patch at best (I would not call them «decent» - they are inherently crude). This is why I insist that they must be overcome with urgency already because they are now here (if a community needed wits and a lunatic appears, wits become more needed). // And no, I am not «crossing»: but people do that, hence I am stating an urgency.

We do not need to simulate the brain, we only ("only") need to implement intelligence. That means the opposite of stating hearsay: it means checking every potential uttering and storing results (and also reinforcing the ways that had the system achieve sophisticated thoughts and conclusions).

It is not given that LLMs cannot be part of such system. They surely have a lot of provisional uttering to criticize.


There's already a lot of research happening for the next thing. The AGI race has been on among not only companies, but also between the largest nations. Everybody's doing their best and then some.

It could very well be that the highest intelligent functionalities requires a closely brain-like substrate. We don't know yet, but we'll get there eventually. And it is very likely to be something emergent, not specifically programmed features as you seem to be insinuating with "... checking [...] and storing results..."


> as you seem to be insinuating with

The implementation details are not clear, not the goals.

I never said that the feature has to be coded explicitly. I said it has to be there.


OK. So it's just a matter of waiting for the desired capabilities to emerge in future models.


But they will probably thought models, not just language models.

The engineering will be different.


Possibly. I personally think it's the type of data and scale that're the primary differentiators. The use of characters is a fundamental flaw because characters are synthetic entities. Instead the models should be based on raw sensory data types, such as pixels and waveforms, and iterate from there on something close to the existing architecture.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: