This is how LLM distorts

With the development of artificial intelligence (AI), more and more attention is being paid to so-called large language models (LLMs), which are now present not only in scientific research but also in many areas of everyday life—for example, in legal work, health data analysis, and computer program coding. However, understanding how these models work remains a serious challenge, especially when they make seemingly inexplicable mistakes or give misleading answers.

A new study by MIT researchers draws attention to a little-known but important phenomenon: so-called positional bias. This bias means that models tend to overemphasize information at the beginning or end of texts, while the middle parts are often pushed into the background. This effect can result, for example, in an AI-based search tool being more likely to find the information you are looking for in a 30-page document if it is on the first or last pages, even if the relevant detail is in the middle.

To explore the root of this phenomenon, the researchers developed a mathematical theoretical framework that examines the functioning of the so-called transformer architecture, which forms the basis of language models. This architecture relies in particular on the so-called attention mechanism, which allows the model to interpret individual words in their textual context. In practice, however, this mechanism has limitations: for example, for the sake of computational efficiency, many models limit how many other words a word can “pay attention” to. One such limitation, causal masking, explicitly favors words at the beginning of a text, even if they are of little significance.

The study also points out that these biases may stem not only from the architecture itself, but also from the data used to train the models. If the training data sets overrepresent information found at the beginning of texts, such a pattern may automatically be incorporated into the model's operation. Position bias is therefore partly a technical issue and partly a data quality issue.

Experiments conducted by researchers have confirmed this phenomenon: when the position of the correct answer in the text was changed in a given task, such as information retrieval, the accuracy of the models decreased dramatically towards the middle of the text and then improved slightly as the answer approached the end of the text. This pattern is referred to in the literature as the “lost in the middle” phenomenon.

Although the problem is not new, the novelty of the study lies in the fact that the researchers identified specific mechanisms that contribute to this bias—and also made suggestions for mitigating it. These include rethinking masking techniques, reducing the number of attention layers, and consciously applying positional encodings, which can help models interpret the text as a whole in a more balanced way.

It is important to emphasize that this phenomenon is not equally problematic in all areas of application. In text composition, for example, it is natural that the beginning and end of a text are given prominence. However, in applications where accurate data extraction or fair decision-making is the goal—such as in legal or medical contexts—these biases can have serious consequences.

Overall, the work of MIT researchers is a step toward making artificial intelligence systems more transparent and reliable. It does not promise an immediate solution, and we should not believe that positional bias is always a serious problem, but it is certain that a better understanding can bring us closer to the responsible and conscious use of AI systems. 

Share this post
Where is Artificial Intelligence Really Today?
The development of artificial intelligence has produced spectacular and often impressive results in recent years. Systems like ChatGPT can generate natural-sounding language, solve problems, and in many tasks, even surpass human performance. However, a growing number of prominent researchers and technology leaders — including John Carmack and François Chollet — caution that these achievements don’t necessarily indicate that artificial general intelligence (AGI) is just around the corner. Behind the impressive performances, new types of challenges and limitations are emerging that go far beyond raw capability.
Rhino Linux Releases New Version: 2025.3
In the world of Linux distributions, two main approaches dominate: on one side, stable systems that are updated infrequently but offer predictability and security; on the other, rolling-release distributions that provide the latest software at the cost of occasional instability. Rhino Linux aims to bridge this divide by combining the up-to-dateness of rolling releases with the stability offered by Ubuntu as its base.
SEAL: The Harbinger of Self-Taught Artificial Intelligence
For years, the dominant belief was that human instruction—through data, labels, fine-tuning, and carefully designed interventions—was the key to advancing artificial intelligence. Today, however, a new paradigm is taking shape. In a recent breakthrough, researchers at MIT introduced SEAL (Self-Adapting Language Models), a system that allows language models to teach themselves. This is not only a technological milestone—it also raises a fundamental question: what role will humans play in the training of intelligent systems in the future?
All it takes is a photo and a voice recording – Alibaba's new artificial intelligence creates a full-body avatar from them
A single voice recording and a photo are enough to create lifelike, full-body virtual characters with facial expressions and emotions – without a studio, actor, or green screen. Alibaba's latest development, an open-source artificial intelligence model called OmniAvatar, promises to do just that. Although the technology is still evolving, it is already worth paying attention to what it enables – and what new questions it raises.
ALT Linux 11.0 Education is the foundation of Russian educational institutions
ALT Linux is a Russian-based Linux distribution built on the RPM package manager, based on the Sisyphus repository. It initially grew out of Russian localization efforts, collaborating with international distributions such as Mandrake and SUSE Linux, with a particular focus on supporting the Cyrillic alphabet.
Spatial intelligence is the next hurdle for AGI to overcome
With the advent of LLM, machines have gained impressive capabilities. What's more, their pace of development has accelerated, with new models appearing every day that make machines even more efficient and give them even better capabilities. However, upon closer inspection, this technology has only just enabled machines to think in one dimension. The world we live in, however, is three-dimensional based on human perception. It is not difficult for a human to determine that something is under or behind a chair, or where a ball flying towards us will land. According to many artificial intelligence researchers, in order for AGI, or artificial general intelligence, to be born, machines must be able to think in three dimensions, and for this, spatial intelligence must be developed.

Linux distribution updates released in the last few days