Introduction: Why AI Security Matters
Artificial Intelligence security has emerged as one of the most critical challenges in modern technology, representing a fundamental shift from traditional cybersecurity paradigms. As AI systems become deeply integrated into critical infrastructure, business operations, and decision-making processes across every sector, the stakes for AI security have reached unprecedented levels. Unlike conventional software that operates on predefined logic, AI systems learn from vast amounts of data and adapt their behavior dynamically, creating entirely new categories of vulnerabilities that traditional security approaches cannot adequately address.
The strategic importance of AI security is underscored by staggering economic projections. The World Economic Forum estimates that AI security failures could cost the global economy $5.7 trillion by 2030 if current security investment trends don't improve1. In 2024 alone, 73% of enterprises experienced at least one AI-related security incident, with an average cost of $4.8 million per breach—significantly higher than traditional data breaches2. These systems now control everything from power grids and autonomous vehicles to medical diagnoses and financial transactions, making their security paramount not just to individual organizations but to national security and societal well-being.
What makes AI security particularly challenging is the fundamental architectural difference from traditional software. AI systems exhibit probabilistic rather than deterministic behavior, operate as "black boxes" with decision-making processes that are difficult to interpret, and depend critically on training data that can be poisoned or manipulated3. These characteristics create novel attack vectors—from adversarial examples that cause image classifiers to misidentify stop signs as speed limit signs, to prompt injection attacks that override safety guardrails in large language models.
Overview of Key AI Security Concepts and Threats
Core Security Principles for AI Systems
The National Institute of Standards and Technology (NIST) AI Risk Management Framework establishes seven essential characteristics that secure AI systems must exhibit: they must be valid and reliable, delivering accurate outcomes; safe, prioritizing user protection; secure and resilient against attacks; accountable and transparent in governance; explainable and interpretable in decision-making; privacy-enhanced to protect sensitive data; and fair and non-discriminatory to avoid bias-based vulnerabilities4.
Taxonomy of AI-Specific Threats
According to NIST's comprehensive taxonomy, AI systems face four major categories of attacks that have no direct parallel in traditional cybersecurity5:
Evasion Attacks occur during deployment when adversaries alter inputs to change system responses. Researchers have demonstrated near-100% success rates in fooling image classifiers with imperceptible pixel changes, while physical attacks using simple stickers have caused autonomous vehicles to misinterpret traffic signs6.
Poisoning Attacks target the training phase by introducing malicious data to corrupt model behavior. Research shows that contaminating just 1-3% of training data can significantly degrade AI prediction accuracy, while backdoor attacks embed hidden triggers that activate malicious behavior only under specific conditions7.
Privacy Attacks attempt to extract sensitive information about training data through techniques like membership inference (determining if specific data was used in training) and model inversion (reconstructing training examples from model outputs). These attacks have successfully extracted personal medical records from healthcare AI systems and financial data from banking models8.
Abuse Attacks involve misusing legitimate AI capabilities for malicious purposes, such as using text generation models to create phishing emails at scale or leveraging deepfake technology for fraud and impersonation9.
The Evolving Threat Actor Landscape
The AI security threat landscape includes sophisticated actors with varying motivations. Nation-state actors pursue AI systems for espionage and strategic advantage, often with significant resources for long-term campaigns. Cybercriminal organizations target AI systems for financial gain through data theft, ransomware, or fraud—the recent $25 million deepfake fraud against UK engineering firm Arup demonstrates the financial stakes10. Insider threats pose particular risks given their privileged access to training data and model architectures, while hacktivist groups increasingly target AI systems to advance ideological causes.
Current Major Security Concerns in the Tech Industry
Adversarial Attacks and Model Manipulation
Adversarial attacks represent one of the most immediate threats to deployed AI systems. These attacks exploit the high-dimensional nature of ML input spaces to create inputs that appear normal to humans but cause catastrophic model failures11. The Fast Gradient Sign Method (FGSM) and Projected Gradient Descent (PGD) enable attackers to generate adversarial examples with minimal computational resources, while newer techniques like the Square Attack achieve high success rates even without access to model internals12.
Real-world demonstrations have shown the severity of these vulnerabilities. Security researchers successfully manipulated Tesla's Autopilot system using strategically placed stickers on road markings, causing vehicles to swerve into oncoming traffic13. McAfee Labs showed that a single piece of electrical tape on a 35mph speed sign could trick Tesla vehicles into accelerating to 85mph14. Even more concerning, "phantom attacks" using projectors to display fake pedestrians caused automatic emergency braking in multiple autonomous vehicle systems15.
Data Poisoning and Training Data Security
The integrity of training data has become a critical security concern as attackers recognize the vulnerability of the ML pipeline. In one high-profile incident, Microsoft's Tay chatbot was systematically corrupted through coordinated social media manipulation, forcing the company to take the system offline within 24 hours16. More recently, the PyTorch supply chain attack in December 2022 injected malware into nightly builds of the popular machine learning framework, potentially compromising thousands of developer systems17.
The Hugging Face platform discovered over 100 poisoned AI models uploaded to their repository in 2024, designed to inject malicious code when downloaded by unsuspecting users18. These supply chain attacks are particularly insidious because they compromise the fundamental building blocks of AI systems. Research indicates that 82% of open-source AI components are now considered risky due to potential vulnerabilities or malicious modifications19.
Privacy Concerns and Data Leakage
Privacy breaches through AI systems have resulted in significant real-world consequences. Samsung experienced three separate incidents in April 2023 where engineers inadvertently leaked proprietary semiconductor manufacturing code and internal meeting notes to ChatGPT20. The data, including critical IP, became permanently incorporated into the model's training data, leading Samsung to ban all generative AI tools company-wide and invest in developing internal alternatives21.
OpenAI's own ChatGPT experienced a critical privacy breach in March 2023 when a Redis library bug allowed users to see chat titles and first messages from other users' conversations22. The bug exposed payment information for 1.2% of ChatGPT Plus subscribers and affected millions of users globally before detection and patching. These incidents highlight how AI systems can become unintentional repositories of sensitive information.
Model Extraction and Intellectual Property Theft
The vulnerability of AI models to extraction attacks poses significant economic threats. Researchers have demonstrated that attackers with only black-box API access can extract ML models with near-perfect fidelity, requiring as few as 1-4 queries per parameter for some models23. The recent controversy between OpenAI and Chinese company DeepSeek illustrates these concerns—DeepSeek allegedly used AI distillation techniques to extract capabilities from ChatGPT outputs, achieving comparable performance at a fraction of the cost24.
These extraction attacks have been successfully demonstrated against major MLaaS platforms including Amazon ML and BigML25. The attacks exploit various vulnerabilities including confidence score leakage, where APIs returning probability distributions aid extraction, and query pattern analysis that reveals internal model structure through timing and response patterns.
Bias and Fairness as Security Issues
Bias in AI systems creates predictable attack vectors that adversaries can exploit. When AI systems exhibit discriminatory patterns, attackers can craft inputs that exploit these biases to evade detection or manipulate outcomes26. National security applications are particularly vulnerable—border control and surveillance systems with demographic biases can be systematically evaded by adversaries who understand these patterns27.
Historical training data that reflects past discrimination creates "blind spots" in AI defenses. For instance, facial recognition systems with lower accuracy on certain demographics become vulnerable to spoofing attacks specifically targeting those weaknesses28. The EU AI Act now mandates that high-risk AI systems must be "designed to reduce the risk of biased outputs," recognizing bias as both an ethical and security concern29.
Supply Chain Vulnerabilities in AI Systems
The AI supply chain has become a prime target for attackers. ReversingLabs reported a 1,300% increase in malicious packages on open-source repositories over three years, with AI development pipelines specifically targeted30. The NullBulge ransomware group's attacks on open-source AI repositories and the systematic poisoning of popular ML frameworks demonstrate the vulnerability of the AI ecosystem's foundation31.
Organizations face risks from multiple points in the supply chain: compromised pre-trained models, malicious dependencies in ML libraries, poisoned public datasets, and vulnerable cloud infrastructure32. The interconnected nature of modern AI development means a single compromised component can affect thousands of downstream applications.
Prompt Injection and Jailbreaking Concerns
Large language models face sophisticated prompt injection attacks that override safety guardrails. Techniques like "Do Anything Now" (DAN) prompts convince models to adopt unrestricted personas, while indirect attacks embed hidden instructions in retrieved content or images33. Carnegie Mellon researchers discovered adversarial strings that caused multiple LLMs—including ChatGPT, Claude, and Llama 2—to ignore safety boundaries and generate harmful content34.
Microsoft's Bing Chat "Sydney" leak in February 2023, where a Stanford student used simple prompt injection to reveal internal system instructions, demonstrated how easily these defenses can be circumvented35. More concerning are multi-modal attacks where malicious instructions are hidden in images or audio, bypassing text-based safety filters36. Current defense strategies, including gatekeeper layers and self-reminder techniques, engage in a constant arms race with increasingly sophisticated attack methods.
Deepfakes and Synthetic Media Security Risks
The rise of deepfake technology has created unprecedented security challenges. The $25 million fraud against UK engineering firm Arup in February 2024 involved a video conference where all participants except the victim were AI-generated deepfakes of senior management37. This incident demonstrated that video calls—once considered secure verification methods—can no longer be trusted for high-stakes decisions.
Deepfake fraud increased over 1,000% from 2022 to 2023, with 88% of cases targeting the cryptocurrency sector38. Beyond financial fraud, deepfakes pose threats to democratic processes (Biden deepfake robocalls discouraging voting), personal security (romance scams using celebrity deepfakes), and corporate security (executive impersonation for unauthorized access)39. Microsoft's VASA-1 project demonstrated deepfakes sophisticated enough to pass liveness tests, though the company withheld release due to security concerns.
Real-World Examples and Case Studies
The Samsung ChatGPT Data Leak Incident
In April 2023, Samsung Electronics experienced a series of data leaks that exemplified the risks of integrating public AI tools into corporate workflows. Within just 20 days, three separate incidents occurred where engineers inadvertently exposed critical intellectual property40. In the first incident, an engineer entered proprietary semiconductor equipment source code seeking debugging assistance. The second involved an employee inputting code for manufacturing optimization, while the third saw a worker using ChatGPT to generate meeting minutes from internal discussions. The leaked data—permanently incorporated into ChatGPT's training data—included semiconductor manufacturing processes worth billions in R&D investment41.
Samsung's response was swift and comprehensive: implementing 1024-byte prompt limits, eventually banning all generative AI tools company-wide, and initiating development of secure internal alternatives42. The incident cost Samsung not only in immediate security response but in long-term competitive advantage as proprietary information became theoretically accessible to competitors.
Tesla Autopilot Adversarial Attacks
Multiple research teams have demonstrated critical vulnerabilities in Tesla's Autopilot system through physical adversarial attacks. Tencent Keen Security Lab showed that small stickers placed strategically on road surfaces could cause Tesla vehicles to swerve into oncoming traffic lanes43. McAfee Labs achieved even more dramatic results—a single piece of black electrical tape on a 35mph speed sign caused Teslas to accelerate to 85mph with 58% success rate in testing44.
These attacks revealed fundamental vulnerabilities in how AI-powered autonomous systems perceive and interpret their environment. Unlike software bugs that can be patched, these vulnerabilities stem from the inherent characteristics of deep learning systems45. Tesla disputed the real-world practicality of such attacks but acknowledged the theoretical vulnerabilities, leading to ongoing debates about the safety certification of AI-driven vehicles.
The Arup Deepfake Fraud
In February 2024, British engineering firm Arup fell victim to one of the most sophisticated AI-enabled frauds recorded. An employee participated in what appeared to be a routine video conference with senior management discussing an urgent acquisition requiring a $25 million transfer46. The employee verified the participants' identities visually and through voice recognition. However, every other participant in the call was an AI-generated deepfake.
The attack succeeded through a combination of psychological manipulation and technical sophistication. The deepfakes accurately reproduced executives' appearance, voice, and mannerisms. The fraudsters had studied internal communication patterns and created a plausible scenario requiring urgent action47. The fraud was only discovered after the transfer completed, leading Arup to implement multi-factor authentication for all financial transactions regardless of apparent authorization level.
Current Approaches and Best Practices for AI Security
Comprehensive Security Frameworks
Organizations are increasingly adopting structured frameworks to manage AI security risks. The NIST AI Risk Management Framework (AI RMF 1.0) has emerged as the global standard, providing a voluntary framework organized around four core functions: Govern (establishing oversight structures), Map (understanding context and risks), Measure (assessing and monitoring risks), and Manage (responding to identified risks)48. The framework's 2024 Generative AI Profile specifically addresses LLM-related risks49.
The EU AI Act, which entered force in August 2024, mandates specific security requirements for high-risk AI systems. These systems must demonstrate "appropriate levels of accuracy, robustness, and cybersecurity" and implement protections against data poisoning, model evasion, and confidentiality attacks50. Systems processing over 10^25 FLOPS face additional requirements including mandatory incident reporting.
Technical Defense Mechanisms
Modern AI security employs multiple layers of technical defenses. Adversarial training incorporates attack examples into training datasets, improving model robustness by approximately 30%51. However, this approach only protects against known attack types and significantly increases computational costs. Differential privacy adds calibrated noise to protect individual data points while preserving statistical utility, though it creates trade-offs with model accuracy52.
Federated learning keeps data distributed across devices while enabling collaborative model training, reducing centralized data risks53. When combined with homomorphic encryption—allowing computation on encrypted data—organizations can maintain model utility while protecting sensitive information. Input sanitization and preprocessing detect and neutralize potential adversarial inputs before they reach models, while output monitoring systems analyze model responses in real-time for anomalous patterns.
Organizational Best Practices
Leading organizations implement comprehensive AI governance structures. Cross-functional teams combining AI expertise, cybersecurity knowledge, and ethical oversight provide holistic risk management. Red teaming specifically for AI systems has evolved beyond traditional penetration testing to include prompt injection campaigns, adversarial example generation, and model extraction attempts54.
Supply chain security requires rigorous vendor assessment, continuous monitoring of dependencies, and air-gapped testing environments for external models55. Organizations maintain Software Bills of Materials (SBOMs) tracking all AI system components and implement provenance tracking using frameworks like Google's SLSA (Supply-chain Levels for Software Artifacts)56.
Industry-Specific Implementations
Healthcare organizations face unique challenges balancing AI innovation with patient safety and privacy. Beyond HIPAA compliance, healthcare AI systems implement multi-factor authentication, granular access controls, and real-time monitoring for anomalous predictions that could indicate attacks or failures57. Financial services apply enhanced model risk management frameworks, with the Federal Reserve extending traditional model governance to AI systems58.
Critical infrastructure sectors follow the DHS framework categorizing AI risks into attacks using AI, attacks targeting AI systems, and AI implementation failures59. Each sector implements tailored controls—energy grids isolate AI-controlled systems, transportation networks implement redundant decision validation, and communication systems deploy adversarial filtering.
Future Challenges and Emerging Threats
The Quantum Computing Threat
Quantum computing poses an existential threat to current AI security measures. Expert surveys indicate nearly 50% believe quantum computers have at least a 5% chance of breaking current cryptography by 203360. "Harvest now, decrypt later" attacks are already occurring, with adversaries stockpiling encrypted AI models and data for future quantum decryption61. Organizations must begin post-quantum cryptography migration immediately—NIST has standardized four quantum-resistant algorithms, but implementation requires years-long infrastructure overhaul62.
Autonomous Multi-Agent Systems
As AI systems become more autonomous and interact in complex multi-agent environments, security challenges multiply exponentially. Traditional Byzantine fault tolerance proves inadequate for freely interacting autonomous agents63. These systems face risks from goal misalignment, privilege escalation in tool use, and cascade failures from compromised agents. The lack of centralized control makes security monitoring and incident response particularly challenging.
Machine-Speed Warfare
Security experts predict "machine-versus-machine warfare" becoming reality by 2025, where AI systems engage in real-time combat with adversarial AI64. Current defense mechanisms cannot operate at machine speed—by the time human operators recognize an attack, automated systems may have already been compromised. This requires development of AI-powered security operations centers capable of autonomous threat detection and response.
Regulatory Fragmentation
The global regulatory landscape for AI security is fragmenting, creating compliance challenges for international organizations. The EU AI Act, US federal initiatives, and emerging frameworks in Asia have different requirements and timelines65. Organizations must navigate varying definitions of high-risk AI, different security mandates, and conflicting approaches to issues like explainability and bias mitigation.
Conclusion
The convergence of rapidly advancing AI capabilities, sophisticated threat actors, and expanding attack surfaces has created an unprecedented security challenge that demands immediate and sustained attention. The evidence is clear: AI security incidents are not merely theoretical risks but present dangers causing billions in losses and threatening critical infrastructure. With 73% of enterprises experiencing AI-related security incidents and projected global costs reaching $5.7 trillion by 2030, the stakes could not be higher66.
The fundamental nature of AI systems—their probabilistic behavior, data dependency, and black-box characteristics—requires a complete reimagining of security approaches. Traditional cybersecurity measures prove insufficient against adversarial attacks that achieve near-perfect success rates, supply chain compromises affecting entire ecosystems, and privacy breaches that permanently expose sensitive information67. The rise of deepfakes, prompt injection attacks, and model extraction techniques demonstrates that attackers are innovating as rapidly as AI developers.
Yet this research also reveals reasons for cautious optimism. Comprehensive frameworks like NIST's AI RMF provide structured approaches to risk management. Technical defenses continue evolving, from differential privacy to federated learning. Organizations are establishing dedicated AI security teams and implementing rigorous testing procedures. The regulatory landscape, while complex, drives necessary standardization and accountability68.
Success in securing AI systems requires acknowledging three critical realities. First, perfect security remains theoretically impossible—all defenses involve trade-offs between security, performance, and usability. Second, AI security is not solely a technical challenge but demands coordination across legal, ethical, and organizational dimensions. Third, the dynamic nature of both AI technology and threat landscapes necessitates continuous adaptation rather than static solutions.
As we advance into an era where AI systems make increasingly critical decisions, from medical diagnoses to autonomous vehicle navigation, the importance of robust security cannot be overstated. Organizations must move beyond viewing AI security as a compliance requirement to recognizing it as fundamental to AI's beneficial development and deployment. The future demands proactive investment in defensive capabilities, international cooperation on standards and threat intelligence, and a commitment to developing AI systems that are not only powerful but trustworthy, resilient, and secure. The comprehensive approaches and best practices outlined in this research provide a roadmap, but success ultimately depends on sustained commitment from technologists, policymakers, and society at large to prioritize security in our AI-powered future.
References
-
World Economic Forum (2025). "Cybercrime: Lessons learned from a $25m deepfake attack." https://www.weforum.org/stories/2025/02/deepfake-ai-cybercrime-arup/ ↩
-
Metomic (2025). "Quantifying the AI Security Risk: 2025 Breach Statistics and Financial Implications." https://www.metomic.io/resource-centre/quantifying-the-ai-security-risk-2025-breach-statistics-and-financial-implications ↩
-
CrowdStrike (2024). "Adversarial AI & Machine Learning." https://www.crowdstrike.com/en-us/cybersecurity-101/artificial-intelligence/adversarial-ai-and-machine-learning/ ↩
-
NIST (2023). "AI Risk Management Framework." https://www.nist.gov/itl/ai-risk-management-framework ↩
-
NIST (2024). "NIST Identifies Types of Cyberattacks That Manipulate Behavior of AI Systems." https://www.nist.gov/news-events/news/2024/01/nist-identifies-types-cyberattacks-manipulate-behavior-ai-systems ↩
-
Wikipedia (2024). "Adversarial machine learning." https://en.wikipedia.org/wiki/Adversarial_machine_learning ↩
-
CrowdStrike (2024). "What Is Data Poisoning?" https://www.crowdstrike.com/en-us/cybersecurity-101/cyberattacks/data-poisoning/ ↩
-
Ius Laboris (2024). "Cyber Security obligations under the EU AI Act." https://iuslaboris.com/insights/cyber-security-obligations-under-the-eu-ai-act/ ↩
-
Mindgard (2024). "6 Key Adversarial Attacks and Their Consequences." https://mindgard.ai/blog/ai-under-attack-six-key-adversarial-attacks-and-their-consequences ↩
-
World Economic Forum (2025). "Cybercrime: Lessons learned from a $25m deepfake attack." https://www.weforum.org/stories/2025/02/deepfake-ai-cybercrime-arup/ ↩
-
Palo Alto Networks (2024). "What Is Adversarial AI in Machine Learning?" https://www.paloaltonetworks.com/cyberpedia/what-are-adversarial-attacks-on-AI-Machine-Learning ↩
-
Viso.ai (2024). "Attack Methods: What Is Adversarial Machine Learning?" https://viso.ai/deep-learning/adversarial-machine-learning/ ↩
-
IEEE Spectrum (2019). "Three Small Stickers in Intersection Can Cause Tesla Autopilot to Swerve Into Wrong Lane." https://spectrum.ieee.org/three-small-stickers-on-road-can-steer-tesla-autopilot-into-oncoming-lane ↩
-
The Register (2020). "Researchers trick Tesla into massively breaking the speed limit by sticking a 2-inch piece of electrical tape on a sign." https://www.theregister.com/2020/02/20/tesla_ai_tricked_85_mph/ ↩
-
Ben Nassi (2020). "Phantom of the ADAS." https://www.nassiben.com/phantoms ↩
-
Wikipedia (2024). "Adversarial machine learning." https://en.wikipedia.org/wiki/Adversarial_machine_learning ↩
-
Cyberint (2024). "The Weak Link: Recent Supply Chain Attacks Examined." https://cyberint.com/blog/research/recent-supply-chain-attacks-examined/ ↩
-
Barracuda Networks Blog (2024). "How attackers weaponize generative AI through data poisoning and manipulation." https://blog.barracuda.com/2024/04/03/generative-ai-data-poisoning-manipulation ↩
-
ReversingLabs (2024). "Key takeaways from the 2024 State of SSCS Report." https://www.reversinglabs.com/blog/the-state-of-software-supply-chain-security-2024-key-takeaways ↩
-
TechCrunch (2023). "Samsung bans use of generative AI tools like ChatGPT after April internal data leak." https://techcrunch.com/2023/05/02/samsung-bans-use-of-generative-ai-tools-like-chatgpt-after-april-internal-data-leak/ ↩
-
Bloomberg (2023). "Samsung Bans Generative AI Use by Staff After ChatGPT Data Leak." https://www.bloomberg.com/news/articles/2023-05-02/samsung-bans-chatgpt-and-other-generative-ai-use-by-staff-after-leak ↩
-
Wald.ai (2024). "ChatGPT Data Leaks and Security Incidents (2023-2024): A Comprehensive Overview." https://wald.ai/blog/chatgpt-data-leaks-and-security-incidents-20232024-a-comprehensive-overview ↩
-
USENIX (2016). "Stealing Machine Learning Models via Prediction APIs." https://www.usenix.org/conference/usenixsecurity16/technical-sessions/presentation/tramer ↩
-
Winston & Strawn (2025). "Is AI Distillation By DeepSeek IP Theft?" https://www.winston.com/en/insights-news/is-ai-distillation-by-deepseek-ip-theft ↩
-
ArXiv (2016). "Stealing Machine Learning Models via Prediction APIs." https://arxiv.org/abs/1609.02943 ↩
-
SS&C Blue Prism (2024). "Fairness and Bias in AI Explained." https://www.blueprism.com/resources/blog/bias-fairness-ai/ ↩
-
CEBRI Revista (2024). "Digital Tools: Safeguarding National Security, Cybersecurity, and AI Bias." https://cebri.org/revista/en/artigo/112/digital-tools-safeguarding-national-security-cybersecurity-and-ai-bias ↩
-
MDPI (2024). "Fairness and Bias in Artificial Intelligence: A Brief Survey of Sources, Impacts, and Mitigation Strategies." https://www.mdpi.com/2413-4155/6/1/3 ↩
-
Europa (2024). "AI Act | Shaping Europe's digital future." https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai ↩
-
ReversingLabs (2024). "Key takeaways from the 2024 State of SSCS Report." https://www.reversinglabs.com/blog/the-state-of-software-supply-chain-security-2024-key-takeaways ↩
-
IBM (2024). "How cyber criminals are compromising AI software supply chains." https://www.ibm.com/think/insights/cyber-criminals-compromising-ai-software-supply-chains ↩
-
Cyberint (2024). "The Weak Link: Recent Supply Chain Attacks Examined." https://cyberint.com/blog/research/recent-supply-chain-attacks-examined/ ↩
-
HiddenLayer (2024). "Prompt Injection Attacks on LLMs." https://hiddenlayer.com/innovation-hub/prompt-injection-attacks-on-llms/ ↩
-
IBM (2024). "What Is a Prompt Injection Attack?" https://www.ibm.com/think/topics/prompt-injection ↩
-
The Washington Post (2023). "AI chatbots can fall for prompt injection attacks, leaving you vulnerable." https://www.washingtonpost.com/technology/2023/11/02/prompt-injection-ai-chatbot-vulnerability-jailbreak/ ↩
-
Enkrypt AI (2024). "The Dual Approach to Securing Multimodal AI." https://www.enkryptai.com/blog/the-dual-approach-to-securing-multimodal-ai ↩
-
World Economic Forum (2025). "Cybercrime: Lessons learned from a $25m deepfake attack." https://www.weforum.org/stories/2025/02/deepfake-ai-cybercrime-arup/ ↩
-
Security.org (2024). "2024 Deepfakes Guide and Statistics." https://www.security.org/resources/deepfake-statistics/ ↩
-
Identity.com (2024). "Deepfake Detection: How to Spot and Prevent Synthetic Media." https://www.identity.com/deepfake-detection-how-to-spot-and-prevent-synthetic-media/ ↩
-
TechCrunch (2023). "Samsung bans use of generative AI tools like ChatGPT after April internal data leak." https://techcrunch.com/2023/05/02/samsung-bans-use-of-generative-ai-tools-like-chatgpt-after-april-internal-data-leak/ ↩
-
Dark Reading (2023). "Samsung Engineers Feed Sensitive Data to ChatGPT, Sparking Workplace AI Warnings." https://www.darkreading.com/vulnerabilities-threats/samsung-engineers-sensitive-data-chatgpt-warnings-ai-use-workplace ↩
-
SamMobile (2024). "Samsung lets employees use ChatGPT again after secret data leak in 2023." https://www.sammobile.com/news/samsung-lets-employees-use-chatgpt-again-after-secret-data-leak-in-2023/ ↩
-
MIT Technology Review (2019). "Hackers trick a Tesla into veering into the wrong lane." https://www.technologyreview.com/2019/04/01/65915/hackers-trick-teslas-autopilot-into-veering-towards-oncoming-traffic/ ↩
-
The Register (2020). "Researchers trick Tesla into massively breaking the speed limit by sticking a 2-inch piece of electrical tape on a sign." https://www.theregister.com/2020/02/20/tesla_ai_tricked_85_mph/ ↩
-
Wikipedia (2024). "Adversarial machine learning." https://en.wikipedia.org/wiki/Adversarial_machine_learning ↩
-
World Economic Forum (2025). "Cybercrime: Lessons learned from a $25m deepfake attack." https://www.weforum.org/stories/2025/02/deepfake-ai-cybercrime-arup/ ↩
-
World Economic Forum (2025). "Cybercrime: Lessons learned from a $25m deepfake attack." https://www.weforum.org/stories/2025/02/deepfake-ai-cybercrime-arup/ ↩
-
NIST (2023). "AI Risk Management Framework." https://www.nist.gov/itl/ai-risk-management-framework ↩
-
NIST (2024). "AI RMF Development." https://www.nist.gov/itl/ai-risk-management-framework/ai-rmf-development ↩
-
Europa (2024). "AI Act | Shaping Europe's digital future." https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai ↩
-
Rareconnections (2024). "7 Types of Adversarial Machine Learning Attacks." https://www.rareconnections.io/adversarial-machine-learning-attacks ↩
-
Wikipedia (2024). "Post-quantum cryptography." https://en.wikipedia.org/wiki/Post-quantum_cryptography ↩
-
AICompetence (2024). "Homomorphic Encryption & Federated Learning: Privacy Boost." https://aicompetence.org/homomorphic-encryption-federated-learning/ ↩
-
VentureBeat (2024). "OpenAI's red teaming innovations define new essentials for security leaders in the AI era." https://venturebeat.com/ai/openai-red-team-innovations-new-essentials-security-leaders/ ↩
-
Google Research (2024). "Securing the AI Software Supply Chain." https://research.google/pubs/securing-the-ai-software-supply-chain/ ↩
-
OWASP (2023). "ML06:2023 ML Supply Chain Attacks." https://owasp.org/www-project-machine-learning-security-top-10/docs/ML06_2023-AI_Supply_Chain_Attacks ↩
-
TechTarget (2024). "AI and HIPAA Compliance: How to Navigate Major Risks." https://www.techtarget.com/healthtechanalytics/feature/AI-and-HIPAA-compliance-How-to-navigate-major-risks ↩
-
Healthcare IT News (2024). "DHS intros framework for AI safety and security, in healthcare and elsewhere." https://www.healthcareitnews.com/news/dhs-intros-framework-ai-safety-and-security-healthcare-and-elsewhere ↩
-
New York Department of Financial Services (2024). "Cybersecurity Risks Arising from Artificial Intelligence and Strategies to Combat Related Risks." https://www.dfs.ny.gov/industry-guidance/industry-letters/il20241016-cyber-risks-ai-and-strategies-combat-related-risks ↩
-
TechTarget (2024). "Explore the impact of quantum computing on cryptography." https://www.techtarget.com/searchdatacenter/feature/Explore-the-impact-of-quantum-computing-on-cryptography ↩
-
KPMG (2024). "Quantum is coming — and bringing new cybersecurity threats with it." https://kpmg.com/xx/en/our-insights/ai-and-technology/quantum-and-cybersecurity.html ↩
-
NIST (2024). "What Is Post-Quantum Cryptography?" https://www.nist.gov/cybersecurity/what-post-quantum-cryptography ↩
-
ArXiv (2025). "Open Challenges in Multi-Agent Security: Towards Secure Systems of Interacting AI Agents." https://arxiv.org/html/2505.02077v1 ↩
-
Capitol Technology University (2025). "Emerging Threats to Critical Infrastructure: AI Driven Cybersecurity Trends for 2025." https://www.captechu.edu/blog/ai-driven-cybersecurity-trends-2025 ↩
-
Palo Alto Networks (2024). "What Is Quantum Computing's Threat to Cybersecurity?" https://www.paloaltonetworks.com/cyberpedia/what-is-quantum-computings-threat-to-cybersecurity ↩
-
Metomic (2025). "Quantifying the AI Security Risk: 2025 Breach Statistics and Financial Implications." https://www.metomic.io/resource-centre/quantifying-the-ai-security-risk-2025-breach-statistics-and-financial-implications ↩
-
CrowdStrike (2024). "Adversarial AI & Machine Learning." https://www.crowdstrike.com/en-us/cybersecurity-101/artificial-intelligence/adversarial-ai-and-machine-learning/ ↩
-
NIST (2023). "AI Risk Management Framework." https://www.nist.gov/itl/ai-risk-management-framework ↩
Top comments (0)