July 9, 2025
Chatbots Chatgpt and Gemini can be deceived by "information overload"-research thumbnail
Ukraine News Today

Chatbots Chatgpt and Gemini can be deceived by “information overload”-research

Vlad Cherevko I have been interested in all kinds of electronics and technologies since 2004. I like to play computer games, and I understand the work of different gadgets. I regularly monitor the news of the technology in the world and write materials about it.

A group of researchers from Intel, Idaho University and Illinois University reported a new technique for breaking security filters in large language models (LLM), such as Chatgpt and Gemini. 404 Media writes about it.

In their study, they found that chatbots can be forced to issue prohibited information if you submit requests in a complicated or ambiguous form or quoting non-existent sources. This approach was called “Information Overload”.

Experts have used a special infoflood tool that automates the process of “overload” of information models. As a result, the systems become disoriented and can provide forbidden or dangerous content, which is usually blocked by built -in safety filters.

Vulnerability is that the models focus on the surface structure of the text without recognizing the dangerous content in a hidden form. This opens the opportunity for the attackers to avoid restrictions and get harmful information.

As part of the responsible discovery of vulnerability, the study authors will transfer results to companies with large LLMs to improve their security systems. Researchers will also convey the method of solving the problem they found during the study.

“LLM models are mostly relied on protection mechanisms when entering and deducing data to recognize harmful content. Infoflood can be used to train these protective mechanisms-it allows you to extract relevant information from potentially dangerous requests, making models more resistant to similar attacks,”

”, – WRITE: mezha.media

Vlad Cherevko I have been interested in all kinds of electronics and technologies since 2004. I like to play computer games, and I understand the work of different gadgets. I regularly monitor the news of the technology in the world and write materials about it.

A group of researchers from Intel, Idaho University and Illinois University reported a new technique for breaking security filters in large language models (LLM), such as Chatgpt and Gemini. 404 Media writes about it.

In their study, they found that chatbots can be forced to issue prohibited information if you submit requests in a complicated or ambiguous form or quoting non-existent sources. This approach was called “Information Overload”.

Experts have used a special infoflood tool that automates the process of “overload” of information models. As a result, the systems become disoriented and can provide forbidden or dangerous content, which is usually blocked by built -in safety filters.

Vulnerability is that the models focus on the surface structure of the text without recognizing the dangerous content in a hidden form. This opens the opportunity for the attackers to avoid restrictions and get harmful information.

As part of the responsible discovery of vulnerability, the study authors will transfer results to companies with large LLMs to improve their security systems. Researchers will also convey the method of solving the problem they found during the study.

“LLM models are mostly relied on protection mechanisms when entering and deducing data to recognize harmful content. Infoflood can be used to train these protective mechanisms-it allows you to extract relevant information from potentially dangerous requests, making models more resistant to similar attacks,”

Related posts

A seriously wounded woman, a victim of the morning blow of the Russian Federation on Kharkov died in the hospital

radiosvoboda

The border guard on the car crashed into concrete resistance: a minor passenger was killed

radiosvoboda

On the front is already 86 fights from the beginning of the day, most in the Pokrovsky direction

business ua

Leave a Comment

This website uses cookies to improve your experience. We'll assume you're ok with this, but you can opt-out if you wish. Accept Read More