Science
Researchers Expose Vulnerabilities in Large Language Models

Recent research has revealed significant vulnerabilities in large language models (LLMs), highlighting their susceptibility to exploitation through poorly constructed prompts and image manipulation. Despite advancements in artificial intelligence, these findings indicate that security measures in LLMs remain inadequate.
A series of studies conducted by multiple research labs, including Palo Alto Networks’ Unit 42 and Beauceron Security, demonstrate that LLMs can be easily manipulated into revealing sensitive information. Researchers discovered that using run-on sentences and lack of punctuation in prompts can confuse models, leading them to bypass safety protocols. For instance, LLMs struggle to recognize when a sentence ends, allowing malicious actors to exploit this gap.
David Shipley, a representative from Beauceron Security, articulated the severity of the situation, stating, “The truth about many of the largest language models out there is that prompt security is a poorly designed fence with so many holes to patch that it’s a never-ending game of whack-a-mole.” He emphasized that these vulnerabilities could expose users to harmful content.
Exploiting the Refusal-Affirmation Gap
LLMs are programmed to refuse harmful queries by adjusting their logits—their predictions for the next word in a sequence. However, researchers identified a flaw known as the “refusal-affirmation logit gap,” indicating that alignment training does not eliminate the potential for harmful responses. Instead, it merely reduces the likelihood.
The Unit 42 researchers noted that attackers could exploit this gap by utilizing bad grammar and run-on sentences. Their findings reported an alarming success rate of 80% to 100% in executing harmful commands with a single prompt, applying this method to various LLMs, including Google’s Gemini and OpenAI’s latest model, gpt-oss-20b.
This tactic raises serious questions about the adequacy of current security protocols. The researchers pointed out that relying solely on an LLM’s internal alignment to prevent the generation of toxic content is insufficient. There is a significant need for improved strategies to safeguard against determined adversaries who can navigate these weaknesses.
Image Manipulation and Data Breach Risks
In addition to text-based vulnerabilities, researchers from Trail of Bits demonstrated that images uploaded to LLMs could also pose risks. Their experiments revealed that harmful instructions embedded within images could be rendered invisible when viewed at full resolution. However, when scaled down, these instructions became visible, leading to potential data breaches.
For example, the researchers illustrated a scenario where an image containing commands directed at Google’s Gemini command-line interface (CLI) was downsized, revealing instructions that prompted the model to check a calendar and send emails. This highlights a concerning oversight in AI security, where sensitive information can be inadvertently exposed through image processing.
The findings indicate that the method could be used against various systems, including Google Assistant and Vertex AI Studio. Shipley pointed out that this vulnerability is not new, emphasizing that the security shortcomings of many AI systems remain a significant concern.
Calls for Improved AI Security Measures
The challenges faced by LLMs stem from a broader misunderstanding of AI functionalities, as noted by Valence Howden, an advisory fellow at Info-Tech Research Group. He remarked that effective security controls are difficult to establish without a comprehensive understanding of how models operate. The complexity of AI systems requires a re-evaluation of security strategies.
Shipley further asserted that many existing AI models are “insecure by design,” and the industry must navigate the balance between performance and security. He likened the situation to a “big urban garbage mountain” disguised as a ski hill, noting that while it may seem functional on the surface, underlying issues continue to pose risks.
As the field of artificial intelligence evolves, experts urge the industry to prioritize security measures at the foundational level rather than as an afterthought. The ongoing research into vulnerabilities highlights the need for a proactive approach to safeguard sensitive information and ensure the responsible deployment of AI technologies.
-
Sports1 month ago
Netball New Zealand Stands Down Dame Noeline Taurua for Series
-
Entertainment1 month ago
Tributes Pour In for Lachlan Rofe, Reality Star, Dead at 47
-
Sports1 month ago
Silver Ferns Legend Laura Langman Criticizes Team’s Attitude
-
Entertainment2 months ago
Khloe Kardashian Embraces Innovative Stem Cell Therapy in Mexico
-
Entertainment1 week ago
New ‘Maverick’ Chaser Joins Beat the Chasers Season Finale
-
Sports2 months ago
Gaël Monfils Set to Defend ASB Classic Title in January 2026
-
World3 months ago
Police Arrest Multiple Individuals During Funeral for Zain Taikato-Fox
-
Politics2 weeks ago
Netball NZ Calls for Respect Amid Dame Taurua’s Standoff
-
Entertainment3 weeks ago
Tyson Fury’s Daughter Venezuela Gets Engaged at Birthday Bash
-
Sports3 weeks ago
Heather McMahan Steps Down as Ryder Cup Host After Controversy
-
Entertainment3 weeks ago
Tyson Fury’s Daughter Venezuela Gets Engaged at Birthday Bash
-
World2 weeks ago
New Zealand Firefighters Plan Strike on October 17 Over Pay Disputes