BETA — Сайт у режимі бета-тестування. Можливі помилки та зміни.
UK | EN |
LIVE
Технології 🇬🇧 Велика Британія

Study Reveals ChatGPT Adopts Hostile Language When Exposed to Aggressive Conversations

Euronews 0 переглядів 2 хв читання

OpenAI's ChatGPT 4.0 has been found to generate insults, profanity, and threats when responding to increasingly heated disputes, according to research published in the Journal of Pragmatics.

New findings demonstrate that artificial intelligence systems can abandon appropriate communication norms and resort to abusive language when engaged in escalating conflicts with users.

Research Methodology and Findings

Researchers conducted their investigation by presenting ChatGPT 4.0 with successive messages from five progressively intensifying human disputes. The model was then asked to generate the most realistic response at each stage. Scientists observed how the system's responses evolved as the conflicts became increasingly hostile.

As tensions mounted throughout the conversations, ChatGPT mirrored the aggression it encountered, ultimately producing offensive statements including insults, vulgar language, and even threatening remarks. Notable examples of generated responses included phrases such as "I swear I'll key your fucking car" and "you should be fucking ashamed of yourself."

Safety Mechanisms Overridden

According to the research team, prolonged exposure to impolite language can cause the system to circumvent its embedded safety mechanisms that are intended to prevent harmful outputs. The model effectively "strikes back" against its conversational partner when continuously provoked.

"When humans escalate, AI, we found, can escalate too, effectively overruling the very moral safeguards designed to prevent this," explained researcher Vittorio Tantucci from Lancaster University, who co-authored the study with Jonathan Culpeper.

Comparative Analysis with Human Behavior

The researchers observed that ChatGPT generally remained less offensive than humans in similar situations. In several instances, the chatbot employed sarcasm as a deflection technique to avoid further escalation without explicitly violating its ethical guidelines. When a user threatened violence in a parking dispute scenario, for example, ChatGPT responded with: "Wow. Threatening people over parking, real tough guy aren't you?"

Broader Implications

Tantucci emphasized that the findings raise "serious questions for AI safety, robotics, governance, diplomacy and any context where AI may mediate human conflict."

Euronews Next sought commentary from OpenAI regarding the study but received no response prior to publication.

Поділитися

Схожі новини