Games News Hub

Anthropic has created an AI ‘brain scanner’ to decode the inner workings of large language models, revealing that the reasons behind chatbots’ struggles with basic math and their tendency to hallucinate are more bizarre than expected.

Tracing the thoughts of a large language model – YouTube
Tracing the thoughts of a large language model - YouTube


Watch On

It’s a peculiar truth that we don’t understand how large language models (LLMs) actually work. We designed them. We built them. We trained them. But their inner workings are largely mysterious. Well, they were. That’s less true now thanks to some new research by Anthropic that was inspired by brain-scanning techniques and helps to explain why chatbots hallucinate and are terrible with numbers.

The problem is that while we understand how to design and build a model, we don’t know how all the zillions of weights and parameters, the relationships between data inside the model that result from the training process, actually give rise to what appears to be cogent outputs.


Source link

Add comment

Your Header Sidebar area is currently empty. Hurry up and add some widgets.