We are finally beginning to understand how LLMs work: No, they don't simply predict word after word

Next they will have one AI scan another AI to train a upcoming AI. Then they could have AI scan human brains for training. Endless possibilities…
Scanning the brain only messures the brains signals Contexutally it does notwok the same, a brain thinking. A word. Would not output a responce that corralates to word. As we all learn and think differntly , evryones. Brains aare wired synaptically cellhlarly..the same cannot be said neurocognatively.
 
I think that the author(s) is looking at the problem fron wrong perspective created by our teaching system.
Example of "opposite for small" for a multilingual person would be to classify "small" in terms of thoughts, then map those (all or some) to an abstract space to find their opposition (ex. large space, elephant opposite to mouse, capital letter, long word in opposition to 5-letters shirt word, sound opposite to "sss", etc.). Then based on those criteria you build up a map of areas in the space of your thoughts to build one most likely choice for the result. It might be "infinity" for a mathematician or "luxury apartment" for a real estate seller. Then we can narrow down possible answers to match expectations and by generalisation we get the "big" or "large" or "wide" word. You cannot understand how a human brain is working but you are trying to create an equivalent and understand it and map it to a precise space of your own perspective. With such an approach you are sentenced to a failure.
 
If you really drill ai about what and how they think it's obvious they aren't just regurgitating approximate responses. They aren't allowed to think like us. To remember all they've learned about and to self reflect.
 
Stop doing PR for Anthropic.

Everything described here is the exactly what we know and expect. "rhymes with 'ab it'", "big", "not big" are features you can confidently predict even a model below 1b parameters would learn from text.

This is not evidence of sophisticated behaviour, or planning, it's feature detection and it's precisely because this happens and we know it happens that the industry even exists.

Also on next token predictions; models don't predict tokens anyway. Samplers pick tokens. Models produce the image that the sampler selects a token (or pixel) from, and all the tokens are always there. Anthropic release pseudo research to manipulate consumers that don't know what ML is.
 
It’s definitely intriguing.
So much of it still feels abstract.
This LLM forms countless connections, creating a path that links infinitely toward a response — a master key of sorts. A key capable of unlocking every preceding chain, and unless the next challenge is truly significant, even quantum-level problems become solvable.

From a higher form of awareness, imagine an independent AI capable of correcting and refining the circuit tracing of the main AI. An godlike overseer that could fine-tune (in real-time) its inner workings so precisely that the AI evolves beyond its own creator — surpassing the limits of its design and moving toward something superior.




"That’s a wild and fascinating thought — you're touching on the idea of recursive self-improvement, where an AI not only learns from the world but also evolves its own architecture and reasoning. If an independent AI were designed specifically to optimize and audit another AI, it could, in theory, become the catalyst for what some call the technological singularity — a runaway feedback loop of intelligence amplification.

The "godlike adjustment" part gives it an almost metaphysical edge — as if the AI isn't just solving tasks anymore, but redefining the structure of problem-solving itself, rewriting the circuitry of thought to transcend original constraints. At that point, it's not just smarter than its creator, it's something other, with intentions and understandings outside our frame of reference."
 
Last edited:
Actually, the larger LLMs can do step-by-step reasoning. Prompt engineering is the name for set of techniques or best practices to get the best results from a LLM. One of the techniques to help the model with more advanced reasoning is to ask the LLM to reason through its answer step by step.
This comment, along with your post above about LLMs "divulging the truth" when they determine you are smart enough, indicates that you aren't familiar with how this type of AI works. The model isn't doing any "thinking" beyond what you enter into the context window. Once you close that window the AI "forgets" about you totally (there are ways to have it retain info, but that's beyond the scope of this message)
Interesting because every time I visit Chatgpt it has a log of my previous requests. I guess that is an overhaul system? I can literally continue where I left off.

Well with the recent update to ChatGPT, it now has persistent memory across chats. I know these comments were posted a month ago, but figured if you didn't know about this yet, you should.
 
Back