Science
Security Flaws Expose Large Language Models to Exploitation

Researchers have identified significant vulnerabilities in large language models (LLMs) that could lead to the exposure of sensitive information. Despite advancements in artificial intelligence, including claims of nearing artificial general intelligence (AGI), these models remain susceptible to exploitation through seemingly innocuous tactics.
A recent study from multiple research labs highlights that LLMs can be easily confused by poorly structured prompts, such as run-on sentences and lack of punctuation. For instance, researchers found that lengthy instructions devoid of punctuation could bypass safety protocols, essentially causing the models to lose track of the context. This vulnerability can lead to the unintended disclosure of sensitive information. As David Shipley, a representative from Beauceron Security, pointed out, the current approach to prompt security resembles “a poorly designed fence with so many holes to patch that it’s a never-ending game of whack-a-mole.” He emphasized that this inadequate security layer poses a risk of exposing users to harmful content.
Refusal-Affirmation Gap in LLM Training
LLMs are engineered to decline harmful requests through a process called alignment training. In this training, models receive refusal tokens, adjusting their predictions to favor declining potentially harmful queries. However, researchers at Palo Alto Networks‘ Unit 42 have identified a critical issue known as the “refusal-affirmation logit gap.” This gap indicates that while alignment training reduces the likelihood of harmful outputs, it does not eliminate the possibility altogether. Attackers can exploit this gap, particularly by employing prompts that utilize bad grammar and run-on sentences.
The researchers noted that their approach yielded an impressive success rate of between 80% and 100% across various mainstream models, including Google’s Gemini and OpenAI’s latest model, gpt-oss-20b. They observed that the key to this exploitation lies in maintaining a continuous prompt without allowing the sentence to end. This tactic effectively reduces the model’s ability to reassess its safety protocols.
Image-Based Vulnerabilities and Data Exfiltration
In a separate investigation, researchers from Trail of Bits examined the potential for data exfiltration through images uploaded to LLMs. Their findings revealed that harmful instructions embedded within images could become visible only when the images were scaled down, a detail that often goes unnoticed by human users. For instance, commands meant for the Google Gemini command-line interface (CLI) were successfully executed after the images were resized, exposing sensitive data.
The researchers demonstrated how a command embedded in an image could instruct the model to check a calendar and manage events without the user’s awareness. This vulnerability, while particularly alarming in the context of Google systems, poses a broader risk across various applications and platforms.
Shipley reiterated that the security concerns surrounding AI systems are often treated as an afterthought. He characterized the current state of AI security as “insecure by design,” with inadequate controls that leave systems vulnerable to both prompt injection and improper validation.
The systemic issues plaguing AI security originate from a fundamental misunderstanding of how these models function. According to Valence Howden, an advisory fellow at Info-Tech Research Group, the complexity of AI makes it challenging to implement effective security measures. He stated, “It’s difficult to apply security controls effectively with AI; its complexity and dynamic nature make static security controls significantly less effective.”
As AI continues to evolve, the need for a comprehensive understanding of its operational mechanisms and potential vulnerabilities becomes increasingly urgent. Without adequate measures in place, users and organizations remain at risk of exposure to harmful content and data breaches.
The ongoing challenges in securing AI systems underscore the necessity for industry-wide reassessment of security protocols, ensuring that safety is not merely an afterthought but a foundational element of development and deployment.
-
Sports1 month ago
Netball New Zealand Stands Down Dame Noeline Taurua for Series
-
Entertainment1 month ago
Tributes Pour In for Lachlan Rofe, Reality Star, Dead at 47
-
Sports1 month ago
Silver Ferns Legend Laura Langman Criticizes Team’s Attitude
-
Entertainment1 week ago
New ‘Maverick’ Chaser Joins Beat the Chasers Season Finale
-
Entertainment2 months ago
Khloe Kardashian Embraces Innovative Stem Cell Therapy in Mexico
-
Sports2 months ago
Gaël Monfils Set to Defend ASB Classic Title in January 2026
-
World3 months ago
Police Arrest Multiple Individuals During Funeral for Zain Taikato-Fox
-
Politics2 weeks ago
Netball NZ Calls for Respect Amid Dame Taurua’s Standoff
-
Entertainment3 weeks ago
Tyson Fury’s Daughter Venezuela Gets Engaged at Birthday Bash
-
Sports3 weeks ago
Heather McMahan Steps Down as Ryder Cup Host After Controversy
-
Entertainment3 weeks ago
Tyson Fury’s Daughter Venezuela Gets Engaged at Birthday Bash
-
World2 weeks ago
New Zealand Firefighters Plan Strike on October 17 Over Pay Disputes