FAST TRACK
See our Fast Start promotion and start your first pentest on The Cobalt Offensive Security Testing Platform for only $4,950.
FAST TRACK
See our Fast Start promotion and start your first pentest on The Cobalt Offensive Security Testing Platform for only $4,950.

Revealing AI Risks in Cybersecurity: Key Insights from the AI Risk Repository

The rise of artificial intelligence (AI) has transformed cybersecurity, offering advanced tools to detect and prevent threats. However, as AI systems become more integrated into security operations, their vulnerabilities also become more apparent. A new AI Risk Repository cataloging failures highlights the critical need for cybersecurity professionals to understand and address these weaknesses to safeguard against emerging risks.

Introduction to AI Risk Repository

The newly launched AI Risk Repository provides a critical resource for cybersecurity professionals, cataloging instances where AI systems have failed in real-world applications. This database serves as a crucial tool for understanding the limitations and risks associated with AI in cybersecurity. 

By tracking these failures, professionals gain insights into how AI can be compromised, where it falls short, and the potential consequences of relying on flawed algorithms. This understanding is essential for developing strategies to mitigate risks and improving the resilience of AI-driven security measures. In an era where teams increasingly rely on AI  to detect and respond to cyber threats, knowing its weaknesses is as important as leveraging its strengths. The AI Risk Repository not only highlights these vulnerabilities but also emphasizes the need for ongoing vigilance and adaptation in the face of evolving AI technologies.

AI Failures in Cybersecurity 

The AI Risk Repository has already cataloged many examples of AI failures and misuse, highlighting the need for vigilance:

  • Evolv's Gun Detection System in Schools: This AI-based weapons detection system, deployed in schools, produced numerous false positives, mistaking everyday school items for weapons. This led to unnecessary manual checks by security personnel, causing significant disruptions and concerns about the reliability of the system.
  • Facial Recognition at Madison Square Garden: The use of facial recognition technology at Madison Square Garden to enforce bans led to legions of lawyers involved in pending litigation with the venue shows the potential of misuse of this technology.

These incidents highlight the potential risks and consequences when AI systems fail in security contexts, emphasizing the importance of careful implementation and ongoing monitoring to prevent such failures.

Types of AI-Related Risks

AI-related risks in cybersecurity are multifaceted, with algorithmic bias, data privacy issues, and AI exploitation by cybercriminals being among the most significant concerns.

  • Algorithmic bias occurs when AI systems inherit prejudices from the data they are trained on, leading to unfair or discriminatory outcomes. In security applications, this could mean biased identification processes or uneven enforcement of security protocols, disproportionately affecting certain groups. Learn more about the risks of AI Data Poisoning and how to prevent this vulnerability in your AI-enabled applications.
  • Data privacy issues arise as AI systems often require vast amounts of data to function effectively. The collection, storage, and processing of this data can expose sensitive information, making it vulnerable to breaches or misuse. Known as Insecure Output Handling, AI systems can also inadvertently leak or misuse personal information, leading to privacy violations.
  • Exploitation by cybercriminals is another pressing risk. AI can be manipulated to perform malicious tasks, such as creating deepfakes, automating phishing attacks, or finding and exploiting vulnerabilities in systems. These exploits can be highly sophisticated and difficult to detect, posing severe threats to security infrastructures.

Leveraging the AI Risk Repository in Cybersecurity

To address AI vulnerabilities and prevent future incidents, cybersecurity experts such as ISACA emphasize an offensive approach that includes continuous monitoring, ethical governance, and rigorous testing.

Using the repository, cybersecurity teams can better understand potential risks such as bias and hallucinations - a significant challenge seen by many security teams in our SANS AI Survey.  The study found that 71% of cybersecurity practitioners using AI in their security processes reported AI systems generating false positives, leading to alert fatigue. 60% of respondents were concerned with AI respecting privacy or creating skewed results from bias. We still have a ways to go as an industry in our understanding of how to use AI within our organizations best. 

The repository serves as a guide for organizations, providing recommendations on how to manage these risks while maximizing the benefits of AI technologies. For example, teams can use the repository to inform their approach to data governance, model training, and performance evaluation, ensuring that AI systems are not only effective but also ethical and secure.

Responsible AI usage necessitates a deep understanding of technology's impact on your processes, systems, and products. AI's rapid evolution often outpaces the security measures to protect these systems, leading to oversights and potential vulnerabilities. To understand these vulnerabilities, AI-embedded systems must be tested by an experienced AI pentester leveraging an offensive approach and social engineering tactics to expose issues and recommend fixes.  

Ideally, teams develop AI with security and privacy built in from the outset, ensuring that systems are designed to handle unexpected or edge-case scenarios effectively. Ethical considerations also play a crucial role in AI security. Establishing ethical guidelines, usage policies, and governance frameworks helps ensure AI is used responsibly, addressing issues like bias and transparency. Testing of the AI systems ensures that these measures used during system construction have the intended outcome. Approaching AI development from both inputs and outputs fosters trust among users and stakeholders, which is essential for successfully deploying AI technologies.

Recommendations to secure AI-enabled applications

To mitigate the risks associated with AI, Cybersecurity teams should implement a series of proactive measures:

  • Regular Audits and Testing: Continuously audit AI models for vulnerabilities by conducting adversarial testing, where AI is exposed to simulated attacks to identify weaknesses.
  • Bias Detection and Mitigation: Incorporate tools and practices that detect and correct algorithmic biases in AI models to ensure fair and accurate decision-making. This includes diverse training data and regular updates to model parameters.
  • Data Privacy Safeguards: Implement robust data privacy practices, such as encryption and anonymization, to protect sensitive information used by AI systems. Ensuring compliance with data protection regulations like GDPR is essential.
  • Ethical AI Governance: Establish ethical guidelines and governance frameworks to guide AI development and deployment. This includes creating clear accountability structures and transparency in AI decision-making processes.
  • Human-in-the-Loop (HITL) Systems: Maintain human oversight in AI-driven processes, especially in critical applications like cybersecurity, to catch potential errors or vulnerabilities that AI may overlook.
  • Collaboration with AI Security Experts: Work closely with AI security experts and continuously update teams on the latest threats, vulnerabilities, and best practices to ensure AI systems remain secure against evolving cyber threats.

Implementing these steps can significantly reduce AI-related risks and enhance the overall security posture of an organization.

Keep safe and compliant by staying diligent 

As AI continues to evolve, its role in cybersecurity presents both opportunities and challenges. The AI Risk Repository highlights the importance of closely monitoring AI failures to protect against emerging risks. By understanding these vulnerabilities, applying expert recommendations, and implementing proactive measures, cybersecurity teams can better safeguard their organizations. Continuous vigilance, adaptation, and collaboration will be key to navigating the complexities of AI in security, ensuring that AI remains a powerful tool for defense rather than a source of vulnerability.

To find out how Cobalt can help you mitigate AI risk, contact us today for your free Cobalt demo.

Back to Blog
About Gisela Hinojosa
Gisela Hinojosa is a Senior Security Consultant at Cobalt with over 5 years of experience as a penetration tester. Gisela performs a wide range of penetration tests including, network, web application, mobile application, Internet of Things (IoT), red teaming, phishing and threat modeling with STRIDE. Gisela currently holds the Security+, GMOB, GPEN and GPWAT certifications. More By Gisela Hinojosa