Science
Security Flaws Expose Large Language Models to Exploitation

Researchers have identified significant vulnerabilities in large language models (LLMs) that could lead to the exposure of sensitive information. Despite advancements in artificial intelligence, including claims of nearing artificial general intelligence (AGI), these models remain susceptible to exploitation through seemingly innocuous tactics.
A recent study from multiple research labs highlights that LLMs can be easily confused by poorly structured prompts, such as run-on sentences and lack of punctuation. For instance, researchers found that lengthy instructions devoid of punctuation could bypass safety protocols, essentially causing the models to lose track of the context. This vulnerability can lead to the unintended disclosure of sensitive information. As David Shipley, a representative from Beauceron Security, pointed out, the current approach to prompt security resembles “a poorly designed fence with so many holes to patch that it’s a never-ending game of whack-a-mole.” He emphasized that this inadequate security layer poses a risk of exposing users to harmful content.
Refusal-Affirmation Gap in LLM Training
LLMs are engineered to decline harmful requests through a process called alignment training. In this training, models receive refusal tokens, adjusting their predictions to favor declining potentially harmful queries. However, researchers at Palo Alto Networks‘ Unit 42 have identified a critical issue known as the “refusal-affirmation logit gap.” This gap indicates that while alignment training reduces the likelihood of harmful outputs, it does not eliminate the possibility altogether. Attackers can exploit this gap, particularly by employing prompts that utilize bad grammar and run-on sentences.
The researchers noted that their approach yielded an impressive success rate of between 80% and 100% across various mainstream models, including Google’s Gemini and OpenAI’s latest model, gpt-oss-20b. They observed that the key to this exploitation lies in maintaining a continuous prompt without allowing the sentence to end. This tactic effectively reduces the model’s ability to reassess its safety protocols.
Image-Based Vulnerabilities and Data Exfiltration
In a separate investigation, researchers from Trail of Bits examined the potential for data exfiltration through images uploaded to LLMs. Their findings revealed that harmful instructions embedded within images could become visible only when the images were scaled down, a detail that often goes unnoticed by human users. For instance, commands meant for the Google Gemini command-line interface (CLI) were successfully executed after the images were resized, exposing sensitive data.
The researchers demonstrated how a command embedded in an image could instruct the model to check a calendar and manage events without the user’s awareness. This vulnerability, while particularly alarming in the context of Google systems, poses a broader risk across various applications and platforms.
Shipley reiterated that the security concerns surrounding AI systems are often treated as an afterthought. He characterized the current state of AI security as “insecure by design,” with inadequate controls that leave systems vulnerable to both prompt injection and improper validation.
The systemic issues plaguing AI security originate from a fundamental misunderstanding of how these models function. According to Valence Howden, an advisory fellow at Info-Tech Research Group, the complexity of AI makes it challenging to implement effective security measures. He stated, “It’s difficult to apply security controls effectively with AI; its complexity and dynamic nature make static security controls significantly less effective.”
As AI continues to evolve, the need for a comprehensive understanding of its operational mechanisms and potential vulnerabilities becomes increasingly urgent. Without adequate measures in place, users and organizations remain at risk of exposure to harmful content and data breaches.
The ongoing challenges in securing AI systems underscore the necessity for industry-wide reassessment of security protocols, ensuring that safety is not merely an afterthought but a foundational element of development and deployment.
-
Sports1 week ago
Gaël Monfils Set to Defend ASB Classic Title in January 2026
-
World4 weeks ago
Police Arrest Multiple Individuals During Funeral for Zain Taikato-Fox
-
Top Stories3 weeks ago
Former Superman Star Dean Cain Joins U.S. Immigration Agency
-
Sports4 weeks ago
Richie Mo’unga’s All Blacks Return Faces Eligibility Hurdles
-
Health4 weeks ago
Navigating the Complexities of ‘Friends with Benefits’ Relationships
-
World4 weeks ago
Fatal ATV Crash Claims Life on Foxton Beach
-
Business3 weeks ago
Grant Taylor Settles Before Zuru Nappy Trial, Shifting Dynamics
-
Sports7 days ago
Warriors Sign Haizyn Mellars on Three-Year Deal Ahead of 2028 Season
-
Entertainment3 weeks ago
Ben MacDonald Exits MasterChef Australia in Fifth Place
-
Entertainment3 weeks ago
New Zealand’s Ben MacDonald Reflects on MasterChef Australia Journey
-
Business1 week ago
Software Glitch Disrupts Air Traffic Control in New Zealand
-
Health3 weeks ago
Qatar Basketball Team Reveals Roster for FIBA Asia Cup 2025