Preview

Philosophical Problems of IT & Cyberspace (PhilIT&C)

Advanced search

Why don’t transformers think like humans?

https://doi.org/10.17726/philIT.2024.2.6

Abstract

Large language models in the form of chatbots very realistically imitate a dialogue as an omniscient interlocutor and therefore have become widespread. But even Google in its Gemini chatbot does not recommend trusting what the chatbot will write and asks to check its answers. In this review, various types of LLM errors such as the curse of inversion, number processing, etc. will be analyzed to identify their causes. Such an analysis led to the conclusion about the common causes of all errors, which is that transformers do not have deep analogy, hierarchy of schemes and selectivity of content taken into account in the inference. But the most important conclusion is that transformers, like other neural networks, are built on the concept of processing the input signal, which creates a strong dependence on superficial noise and irrelevant information that the transformer’s attention layer cannot compensate for. The concept of neural networks was laid down in the 1950s by the idea of F. Rosenblatt’s perceptron and did not take into account the achievements of cognitive psychology that appeared later. According to the constructivist paradigm, the input word (or perception) is only a way to check the correctness of the constructed predictive model for possible situations. This is the cause of the biggest problem of transformers, called hallucinations. And its elimination is possible only by changing the architecture of the neural network, but not by increasing the amount of data in training.

About the Author

A. B. Khomyakov

Russian Federation

Alexander B. Khomyakov, Master of Physical Sciences

Saint-Petersburg 



References

1. Robison K. OpenAI cofounder Ilya Sutskever says the way AI is built is about to change // The Verge, Dec 14, 2024. https://www.theverge.com/2024/12/13/24320811/what-ilya-sutskever-sees-openai-model-datatraining.

2. Neisser W. Cognition and Reality. M.: Progress, 1981. P. 42-43. 230 p.

3. Mitchell M. How do we know how smart ai systems are? // Science, 381(6654). https://www.science.org/doi/10.1126/science.adj5957.

4. Nezhurina M. Alice in Wonderland: Simple Tasks Showing Complete Reasoning Breakdown in State-Of-the-Art Large Language Models. https://arxiv.org/html/2406.02061v1#bib.bib12, 04 Jun 2024.

5. Luria A. R. Fundamentals of Neuropsychology: textbook. M.: Publishing Center “Academy”, 2003. 384 p., p. 123-126.

6. Berglund L. and etc. The Reversal Curse: LLMs trained on “A is B” fail to learn “B is A”. https://arxiv.org/abs/2309.12288, 26 May 2024.

7. Hagag B. Discover What Every Neuron in the Llama Model Does // Towards Data Science. https://towardsdatascience.com/-0927524e4807, Oct 25, 2024.

8. Khomyakov A. B., Chizhik P. A new way of finding analogues as an opportunity to study language, thinking and build artificial intelligence systems // Philosophical Problems of IT & Cyberspace (PhilIT&C). 2024. № 1. Р. 77-88. (In Russ.). https://doi.org/10.17726/philIT.2024.1.5.

9. De La Cruz R. Frank Rosenblatt’s Perceptron, Birth of The Neural Network // Medium. https://medium.com/@robdelacruz/frank-rosenblattsperceptron-19fcce9d627f, 1 November 2023.

10. Tiehen J. Perception as controlled hallucination // ResearchGate.org. https://www.researchgate.net/publication/359601789_Perception_as_controlled_hallucination march 2022.

11. Hoffman D. How Our Senses Deceive Us. M.: AST, 2022.


Review

For citations:


Khomyakov A.B. Why don’t transformers think like humans? Philosophical Problems of IT & Cyberspace (PhilIT&C). 2024;(2):87-98. (In Russ.) https://doi.org/10.17726/philIT.2024.2.6

Views: 208


Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.


ISSN 2305-3763 (Online)