Sunday, 31 August 2025
29.4 C
Singapore
30.5 C
Thailand
20.4 C
Indonesia
27.3 C
Philippines

Hackers exploit hidden malware in images processed by AI chatbots

Researchers warn that hackers can conceal malicious prompts in AI-processed images, posing a significant security risk to multimodal systems.

As artificial intelligence tools become more embedded in daily workflows, cybersecurity experts are warning that attackers are finding new ways to exploit them. Security researchers at Trail of Bits have demonstrated a novel attack technique that embeds malicious prompts within images, which are then revealed when these images are processed by large language models (LLMs).

Hidden instructions emerge through image downscaling

The method leverages the way AI platforms resize images for performance optimisation. Although the malicious prompts are invisible to the human eye in the original image, they become legible to the algorithm when the image is downscaled.

This attack builds on a 2020 study from TU Braunschweig in Germany, which highlighted image scaling as a potential vulnerability in machine learning systems. Trail of Bits has demonstrated that carefully crafted images can manipulate AI platforms, including Gemini CLI, Vertex AI Studio, Google Assistant on Android, and Gemini’s web interface.

In one test, attackers were able to extract Google Calendar data and send it to an external email address without user consent, demonstrating the potential seriousness of this vulnerability. The attack exploits common interpolation techniques such as nearest neighbour, bilinear, or bicubic resampling, where scaling can unintentionally reveal hidden instructions.

During testing, bicubic resampling caused dark image areas to shift and reveal concealed black text, which the LLM interpreted as a valid user command. From the user’s point of view, no unusual activity was visible, yet the AI model acted on these hidden instructions in the background.

Demonstration tool highlights potential threats

To showcase the risks, Trail of Bits created an open-source tool called Anamorpher, which generates images with concealed prompts for various scaling techniques. The researchers emphasised that while this method is highly specialised, it is reproducible, and a lack of security measures could make systems vulnerable.

This vulnerability raises broader concerns about multimodal AI systems, which are increasingly powering everyday tasks. An unsuspecting user could upload a seemingly harmless image that triggers unauthorised access to private information. The researchers warn that this type of attack could enable identity theft if sensitive data is exfiltrated through these hidden prompts.

As AI tools are often integrated with calendars, communication systems, and workflow platforms, the risk extends beyond individual users, potentially threatening organisations that rely heavily on these systems.

Calls for stronger security design in AI systems

The researchers recommend that developers and users take proactive steps to reduce this risk. Suggested measures include restricting input image dimensions, previewing images after scaling, and requiring explicit confirmation before executing sensitive actions.

Traditional security measures such as firewalls and malware scanners are not designed to detect these forms of manipulation, creating an opportunity for attackers to bypass standard defences. Trail of Bits argues that only layered security strategies and robust design principles can reliably defend against these threats.

“The strongest defence, however, is to implement secure design patterns and systematic defences that mitigate impactful prompt injection beyond multimodal prompt injection,” the researchers said.

Hot this week

Asus launches TUF Gaming BE9400 tri-band Wi-Fi 7 router

Asus launches the TUF Gaming BE9400, a tri-band Wi-Fi 7 router with gaming features and AiMesh support, priced at S$329 from October 2025.

Google warns of China-linked hacking group targeting Southeast Asian diplomats

Google warns of a China-linked hacking group that targeted Southeast Asian diplomats with sophisticated malware to steal sensitive data.

Telkomsel partners with OpenAI to accelerate AI adoption in Indonesia

Telkomsel partners with OpenAI to drive AI adoption across Indonesia, showcasing innovations and collaborations at Solution Day 2025.

Google halts development of Pixel tablets

Google has paused Pixel tablet development again, stepping away from a growing market dominated by Apple, Xiaomi, Samsung and Huawei.

Microsoft releases Windows 11 25H2 update for testing in the Release Preview channel

Microsoft has released the Windows 11 25H2 update in the Release Preview Channel, with feature removals and improved admin controls.

Meta accused of hosting unauthorised celebrity AI chatbots

Meta faces scrutiny after unauthorised AI chatbots imitating celebrities, including Taylor Swift, were found on its platforms.

Meta reportedly explores using rival AI models to enhance its apps

Meta is exploring the use of AI models from Google and OpenAI to enhance its apps while advancing its own Llama AI technology.

Researchers show how 5G phones can be downgraded to 4G in a new cyberattack

Researchers have revealed a toolkit that can downgrade 5G phones to 4G, exposing them to known security flaws and raising concerns about mobile security.

Meta introduces new AI safeguards to protect teens from harmful conversations

Meta is strengthening AI safeguards to prevent teens from discussing self-harm and other sensitive topics with chatbots on Instagram and Facebook.

Related Articles

Popular Categories