Science
Researchers Expose Vulnerabilities in Large Language Models

Recent research has revealed significant vulnerabilities in large language models (LLMs), highlighting their susceptibility to exploitation through poorly constructed prompts and image manipulation. Despite advancements in artificial intelligence, these findings indicate that security measures in LLMs remain inadequate.
A series of studies conducted by multiple research labs, including Palo Alto Networks’ Unit 42 and Beauceron Security, demonstrate that LLMs can be easily manipulated into revealing sensitive information. Researchers discovered that using run-on sentences and lack of punctuation in prompts can confuse models, leading them to bypass safety protocols. For instance, LLMs struggle to recognize when a sentence ends, allowing malicious actors to exploit this gap.
David Shipley, a representative from Beauceron Security, articulated the severity of the situation, stating, “The truth about many of the largest language models out there is that prompt security is a poorly designed fence with so many holes to patch that it’s a never-ending game of whack-a-mole.” He emphasized that these vulnerabilities could expose users to harmful content.
Exploiting the Refusal-Affirmation Gap
LLMs are programmed to refuse harmful queries by adjusting their logits—their predictions for the next word in a sequence. However, researchers identified a flaw known as the “refusal-affirmation logit gap,” indicating that alignment training does not eliminate the potential for harmful responses. Instead, it merely reduces the likelihood.
The Unit 42 researchers noted that attackers could exploit this gap by utilizing bad grammar and run-on sentences. Their findings reported an alarming success rate of 80% to 100% in executing harmful commands with a single prompt, applying this method to various LLMs, including Google’s Gemini and OpenAI’s latest model, gpt-oss-20b.
This tactic raises serious questions about the adequacy of current security protocols. The researchers pointed out that relying solely on an LLM’s internal alignment to prevent the generation of toxic content is insufficient. There is a significant need for improved strategies to safeguard against determined adversaries who can navigate these weaknesses.
Image Manipulation and Data Breach Risks
In addition to text-based vulnerabilities, researchers from Trail of Bits demonstrated that images uploaded to LLMs could also pose risks. Their experiments revealed that harmful instructions embedded within images could be rendered invisible when viewed at full resolution. However, when scaled down, these instructions became visible, leading to potential data breaches.
For example, the researchers illustrated a scenario where an image containing commands directed at Google’s Gemini command-line interface (CLI) was downsized, revealing instructions that prompted the model to check a calendar and send emails. This highlights a concerning oversight in AI security, where sensitive information can be inadvertently exposed through image processing.
The findings indicate that the method could be used against various systems, including Google Assistant and Vertex AI Studio. Shipley pointed out that this vulnerability is not new, emphasizing that the security shortcomings of many AI systems remain a significant concern.
Calls for Improved AI Security Measures
The challenges faced by LLMs stem from a broader misunderstanding of AI functionalities, as noted by Valence Howden, an advisory fellow at Info-Tech Research Group. He remarked that effective security controls are difficult to establish without a comprehensive understanding of how models operate. The complexity of AI systems requires a re-evaluation of security strategies.
Shipley further asserted that many existing AI models are “insecure by design,” and the industry must navigate the balance between performance and security. He likened the situation to a “big urban garbage mountain” disguised as a ski hill, noting that while it may seem functional on the surface, underlying issues continue to pose risks.
As the field of artificial intelligence evolves, experts urge the industry to prioritize security measures at the foundational level rather than as an afterthought. The ongoing research into vulnerabilities highlights the need for a proactive approach to safeguard sensitive information and ensure the responsible deployment of AI technologies.
-
Sports1 week ago
Gaël Monfils Set to Defend ASB Classic Title in January 2026
-
World4 weeks ago
Police Arrest Multiple Individuals During Funeral for Zain Taikato-Fox
-
Top Stories3 weeks ago
Former Superman Star Dean Cain Joins U.S. Immigration Agency
-
Sports4 weeks ago
Richie Mo’unga’s All Blacks Return Faces Eligibility Hurdles
-
Health4 weeks ago
Navigating the Complexities of ‘Friends with Benefits’ Relationships
-
World4 weeks ago
Fatal ATV Crash Claims Life on Foxton Beach
-
Business3 weeks ago
Grant Taylor Settles Before Zuru Nappy Trial, Shifting Dynamics
-
Sports1 week ago
Warriors Sign Haizyn Mellars on Three-Year Deal Ahead of 2028 Season
-
Entertainment3 weeks ago
Ben MacDonald Exits MasterChef Australia in Fifth Place
-
Entertainment3 weeks ago
New Zealand’s Ben MacDonald Reflects on MasterChef Australia Journey
-
Business1 week ago
Software Glitch Disrupts Air Traffic Control in New Zealand
-
Health4 weeks ago
Qatar Basketball Team Reveals Roster for FIBA Asia Cup 2025