DarkMind A Novel Backdoor Attack Exploits Reasoning Capabilities of Customized LLMs

While the security analysts noted that unlike conventional backdoor attacks that rely on poisoned training data or overt triggers in user prompts, DarkMind embeds latent triggers directly into the model’s reasoning chain. Dubbed DarkMind, this backdoor attack exploits the reasoning capabilities of LLMs to covertly manipulate outputs without requiring direct user query manipulation. Cyber Security News is a Dedicated News Platform For Cyber News, Cyber Attack News, Hacking News & Vulnerability Analysis. Instant Triggers (τIns): Modify subsequent reasoning steps immediately upon activation (replacing correct arithmetic operators with incorrect ones). The attack raises critical concerns about the security of AI agents deployed across platforms like OpenAI’s GPT Store, which hosts over 3 million customized models. These triggers activate during intermediate processing steps, dynamically altering the final output while leaving the model’s surface-level behavior intact. Retrospective Triggers (τRet): Append malicious reasoning steps after initial processing to reverse or distort conclusions. With years of experience under his belt in Cyber Security, he is covering Cyber Security News, technology and other news. A groundbreaking study by researchers Zhen Guo and Reza Tourani at Saint Louis University has exposed a novel vulnerability in customized large language models (LLMs) like GPT-4o and LLaMA-3. The attack modifies intermediate CoT steps while maintaining plausible final outputs, rendering detection through output monitoring nearly impossible. DarkMind targets the Chain-of-Thought (CoT) reasoning process—the step-by-step logic LLMs use to solve complex tasks. The researchers tested DarkMind across eight datasets spanning arithmetic, commonsense, and symbolic reasoning tasks.

This Cyber News was published on cybersecuritynews.com. Publication date: Tue, 18 Feb 2025 11:05:13 +0000


Cyber News related to DarkMind A Novel Backdoor Attack Exploits Reasoning Capabilities of Customized LLMs

DarkMind A Novel Backdoor Attack Exploits Reasoning Capabilities of Customized LLMs - While the security analysts noted that unlike conventional backdoor attacks that rely on poisoned training data or overt triggers in user prompts, DarkMind embeds latent triggers directly into the model’s reasoning chain. Dubbed DarkMind, this ...
2 months ago Cybersecuritynews.com
The age of weaponized LLMs is here - It's exactly what one researcher, Julian Hazell, was able to simulate, adding to a collection of studies that, altogether, signify a seismic shift in cyber threats: the era of weaponized LLMs is here. The research all adds up to one thing: LLMs are ...
1 year ago Venturebeat.com
Exploring the Security Risks of LLM - According to a recent survey, 74% of IT decision-makers have expressed concerns about the cybersecurity risks associated with LLMs, such as the potential for spreading misinformation. Security Concerns of LLMs While the potential applications of ...
1 year ago Feeds.dzone.com
The impact of prompt injection in LLM agents - This risk is particularly alarming when LLMs are turned into agents that interact directly with the external world, utilizing tools to fetch data or execute actions. Malicious actors can leverage prompt injection techniques to generate unintended and ...
1 year ago Helpnetsecurity.com
BianLian GOs for PowerShell After TeamCity Exploitation - In conjunction with GuidePoint's DFIR team, we responded to an incident that began with the exploitation of a TeamCity server which resulted in the deployment of a PowerShell implementation of BianLian's GO backdoor. The threat actor identified a ...
1 year ago Securityboulevard.com CVE-2024-27198 CVE-2023-42793 BianLian
DeepSeek-R1 Prompts Exploited to Create Sophisticated Malware & Phishing Pages - Trend Micro researchers noted that these vulnerabilities can be weaponized through carefully crafted prompt attacks, where malicious actors design inputs specifically to achieve objectives like jailbreaking the model, extracting sensitive ...
2 weeks ago Cybersecuritynews.com
DARPA awards $1 million to Trail of Bits for AI Cyber Challenge - We're excited to share that Trail of Bits has been selected as one of the seven exclusive teams to participate in the small business track for DARPA's AI Cyber Challenge. Our team will receive a $1 million award to create a Cyber Reasoning System and ...
1 year ago Securityboulevard.com
Integrating LLMs into security operations using Wazuh - Once YARA identifies a malicious file, ChatGPT enriches the alert with details about the detected threat, helping security teams better understand and respond to the incident. Log analysis and data enrichment: Trained LLMs like ChatGPT can interpret ...
2 months ago Bleepingcomputer.com
Russian Sandworm Group Using Novel Backdoor to Target Ukraine - Russian nation-state group Sandworm is believed to be utilizing a novel backdoor to target organizations in Ukraine and other Eastern and Central European countries, according to WithSecure researchers. The previously unreported backdoor, dubbed ...
1 year ago Infosecurity-magazine.com
Why training LLMs with endpoint data will strengthen cybersecurity - Capturing weak signals across endpoints and predicting potential intrusion attempt patterns is a perfect challenge for Large Language Models to take on. The goal is to mine attack data to find new threat patterns and correlations while fine-tuning ...
1 year ago Venturebeat.com
OWASP Top 10 for LLM Applications: A Quick Guide - Even still, the expertise and insights provided, including prevention and mitigation techniques, are highly valuable to anyone building or interfacing with LLM applications. Prompt injections are maliciously crafted inputs that lead to an LLM ...
1 year ago Securityboulevard.com
ChatGPT's o4-mini, o4-mini-high and o3 spotted ahead of release - According to references found on ChatGPT's web app, OpenAI is preparing to launch as many as three new models: o3 (a full-fledged reasoning model in the o3-mini, o3-mini-high lineup), and o4-mini, o4-mini-high, similar to the existing models but with ...
1 week ago Bleepingcomputer.com
4 key devsecops skills for the generative AI era - Experts believe that generative AI capabilities, copilots, and large language models are ushering in a new era of how developers, data scientists, and engineers will work and innovate. They expect AI to improve productivity, quality, and innovation, ...
1 year ago Infoworld.com
9 Best DDoS Protection Service Providers for 2024 - eSecurity Planet content and product recommendations are editorially independent. We may make money when you click on links to our partners. Learn More. One of the most powerful defenses an organization can employ against distributed ...
1 year ago Esecurityplanet.com
Akto Launches Proactive GenAI Security Testing Solution - With the increasing reliance on GenAI models and Language Learning Models like ChatGPT, the need for robust security measures have become paramount. Akto, a leading API Security company, is proud to announce the launch of its revolutionary GenAI ...
1 year ago Darkreading.com
LLMs Open to Manipulation Using Doctored Images, Audio - Such attacks could become a major issue as LLMs become increasingly multimodal or are capable of responding contextually to inputs that combine text, audio, pictures, and even video. Hiding Instructions in Images and Audio At Black Hat Europe 2023 ...
1 year ago Darkreading.com
Researchers Show How to Use One LLM to Jailbreak Another - The exploding use of large language models in industry and across organizations has sparked a flurry of research activity focused on testing the susceptibility of LLMs to generate harmful and biased content when prompted in specific ways. The latest ...
1 year ago Darkreading.com
New 'SpectralBlur' macOS Backdoor Linked to North Korea - Security researchers have dived into the inner workings of SpectralBlur, a new macOS backdoor that appears linked to the recently identified North Korean malware family KandyKorn. The observed SpectralBlur sample was initially uploaded to VirusTotal ...
1 year ago Securityweek.com
Cybercriminals are Showing Hesitation to Utilize AI Cyber Attacks - Media reports highlight the sale of LLMs like WormGPT and FraudGPT on underground forums. Fears mount over their potential for creating mutating malware, fueling a craze in the cybercriminal underground. Concerns arise over the dual-use nature of ...
1 year ago Cybersecuritynews.com
Attack Vector vs Attack Surface: The Subtle Difference - Cybersecurity discussions about "Attack vectors" and "Attack surfaces" sometimes use these two terms interchangeably. This article guides you through the distinctions between attack vectors and attack surfaces to help you better understand the two ...
2 years ago Trendmicro.com
New Sophisticated Linux-Backdoor Attacking OT Systems Exploiting 0-Day RCE - OrpaCrab exemplifies how attackers can compromise critical infrastructure without deep knowledge of industrial protocols, instead leveraging common networking standards to hide malicious traffic within legitimate communications. A sophisticated ...
3 weeks ago Cybersecuritynews.com
An Argument for Coordinated Disclosure of New Exploits - There were more than 23,000 vulnerabilities discovered and disclosed. While not all of them had associated exploits, it has become more and more common for there to be a proverbial race to the bottom to see who can be the first to release an exploit ...
10 months ago Darkreading.com
OpenAI details ChatGPT-o3, o4-mini, o4-mini-high usage limits - OpenAI has launched three new reasoning models - o3, o4-mini, and o4-mini-high for Plus and Pro subscribers, but as it turns out, these models do not offer 'unlimited' usage. In a support document, OpenAI shed light on how you can use ChatGPT's ...
4 days ago Bleepingcomputer.com
Raspberry Robin malware evolves with early access to Windows exploits - Recent versions of the Raspberry Robin malware are stealthier and implement one-day exploits that are deployed only on systems that are susceptible to them. One-day exploits refer to code that leverages a vulnerability that the developer of the ...
1 year ago Bleepingcomputer.com CVE-2023-36802 CVE-2023-29360
New FamousSparrow Malware Attacking Hotels & Engineering Companies to New Backdoor - Most notably, researchers found that one version of SparrowDoor is now modular, while another resembles what other security firms have called “CrowDoor” and attributed to the Earth Estries APT group, suggesting potential overlaps between ...
3 weeks ago Cybersecuritynews.com

Latest Cyber News


Cyber Trends (last 7 days)