EVE Security | Agent AI Security

Was Your AI Agent Hacked?

Or Made a Decision You Can’t Explain?

How CISOs Can Secure Agentic AI Using OWASP, FINOS, and NIST Together


The question is no longer “Is our AI secure?”

The real question is
“Do we understand what our agents are doing right now, why they are doing it, and can we stop them if they go wrong?”

Agentic AI systems plan, act, invoke tools, manage memory, and delegate work to other agents. When something fails, it is often not a classic vulnerability exploit.

It is an autonomous decision that crossed a boundary no one instrumented, governed, or measured.

This is where traditional application security, and even early AI governance programs, fall short.


Three Frameworks Now Define the Path Forward

Three authoritative publications now define how organizations should approach this problem, each from a different and complementary angle:

  1. OWASP Top 10 for Agentic Applications (2026) identifies agent-specific failure modes and provides technical mitigation guidance.
  2. FINOS AI Governance Framework (AIR v2) translates AI risk into enterprise-grade controls, ownership, and auditability.
  3. NIST AI Risk Management Framework (AI RMF 1.0) provides the management model for governing and measuring AI risk over time.

Used together, they form a complete and defensible approach to securing agentic AI in production.


What Each Framework Is Designed to Do


OWASP Top 10 for Agentic Applications (2026)

Threats and technical mitigations

The OWASP Top 10 for Agentic Applications is a security threat taxonomy with mitigation guidance. It focuses on how agentic systems fail in practice and what defenders can do to reduce those risks.

It introduces agent-specific risks such as:

  • ASI-01 Agent Goal Hijack
  • ASI-02 Tool Misuse and Exploitation

  • ASI-03 Identity and Privilege Abuse

  • ASI-06 Memory and Context Poisoning

  • ASI-07 Insecure Inter-Agent Communication

  • ASI-08 Cascading Failures

  • ASI-10 Rogue Agents

For each category, OWASP includes examples, impact analysis, and technical mitigations such as isolation, sandboxing, least-privilege execution, approval gates, and logging.

OWASP’s strength is showing what breaks and how to defend against it at the system and architecture level.

Source: OWASP Top 10 for Agentic Applications (2026), OWASP GenAI Project
https://genai.owasp.org/


FINOS AI Governance Framework (AIR v2)

Controls and accountability

The FINOS AI Governance Framework is an enterprise governance and risk framework. It focuses on how organizations should approve, operate, and oversee AI systems across operational, security, and regulatory dimensions.

Key characteristics:

  • Explicit AI risk catalog spanning operational, cybersecurity, data, and regulatory risk

  • Defined mitigations mapped to those risks

  • Emphasis on observability, accountability, auditability, and resilience

  • Designed to meet regulated-industry expectations, especially financial services

FINOS mitigations are governance-grade. They define what controls must exist, who owns them, and how they support assurance and compliance.

Source: FINOS AI Governance Framework, AIR v2
https://air-governance-framework.finos.org/


NIST AI Risk Management Framework (AI RMF 1.0)

Risk management and measurement

NIST AI RMF provides a cross-sector, outcome-based risk management model for AI systems. It is voluntary, regulator-neutral, and designed to integrate with existing enterprise risk and cybersecurity programs.

The framework is organized around four core functions:

  • Govern – establish accountability, oversight, and policies

  • Map – understand context, purpose, and risk exposure

  • Measure – assess and track risk outcomes

  • Manage – prioritize and treat risk over time

NIST does not define agent-specific threats or prescribe technical controls. Its value is ensuring organizations define acceptable risk, measure whether controls are effective, and continuously adjust as systems evolve.

Source: NIST AI Risk Management Framework 1.0
https://www.nist.gov/itl/ai-risk-management-framework


Where OWASP, FINOS, and NIST Align on Agentic AI Risk


Autonomy increases blast radius and must be constrained

OWASP identifies autonomy as a core risk multiplier and provides mitigations such as execution limits, approval gates, and constrained planning.

FINOS complements this with governance controls like:

  • Agent authority least-privilege frameworks

  • Defined escalation and approval paths

  • Clear operational boundaries for AI systems

NIST reinforces this by requiring organizations to map autonomy to risk tolerance and measure whether outcomes remain acceptable over time.

FINOS reference: Mitigation MI-18, Agent authority least privilege framework
https://air-governance-framework.finos.org/


Tools and integrations are a primary attack surface

OWASP elevates tool misuse and integration abuse as a top agentic risk and recommends mitigations such as allowlists, isolation, and execution controls.

FINOS provides enterprise-level mitigations for:

  • Tool chain validation and sanitization

  • Governance of third-party services and dependencies

  • Control of agent permissions at runtime

NIST adds the expectation that these risks are not just approved once, but measured and managed continuously.

FINOS reference: Mitigation MI-19, Tool chain validation and sanitization
https://air-governance-framework.finos.org/


Identity, privilege, and secrets are central to agent security

OWASP explicitly calls out identity and privilege abuse in agentic systems and provides technical mitigation guidance.

FINOS treats agents as privileged non-human actors and defines mitigations to:

  • Protect credentials and secrets

  • Enforce scoped and revocable permissions

  • Prevent credential discovery or exfiltration

NIST frames this as an accountability and security outcome that must be monitored throughout the system lifecycle.

FINOS reference: Mitigation MI-23, Agentic system credential protection framework
https://air-governance-framework.finos.org/


Observability is required to make mitigations effective

OWASP highlights insufficient logging and observability as a systemic failure in agentic systems and recommends detailed runtime visibility.

FINOS formalizes this through explicit mitigations:

  • AI system observability

  • Agent decision audit and explainability

NIST reinforces that without observability, organizations cannot measure risk or demonstrate control effectiveness.

FINOS references: MI-4, AI system observability · MI-21, Agent decision audit and explainability
https://air-governance-framework.finos.org/


Multi-agent systems fail through cascades

OWASP introduces risks related to insecure inter-agent communication and cascading failures, with mitigations focused on isolation and containment.

FINOS explicitly models multi-agent trust boundary violations and provides governance-grade mitigations for segmentation and isolation.

NIST frames cascading failures as systemic risk that must be understood, measured, and governed at the organizational level.

FINOS references: Risk RI-28, Multi-agent trust boundary violations · Mitigation MI-22, Multi-agent isolation and segmentation
https://air-governance-framework.finos.org/


How Security Teams Should Use All Three Frameworks Together

A practical operating model looks like this:

  • Use OWASP to identify agentic threats and apply technical mitigations at the system and architecture level.

  • Use FINOS to ensure those mitigations become governed, owned, auditable controls.

  • Use NIST to measure effectiveness, manage residual risk, and adapt decisions as agent autonomy and scope evolve.

Put simply:

OWASP defines threats and technical mitigations
FINOS defines enterprise controls and accountability
NIST defines risk governance and measurement


Final Takeaway

OWASP shows how agentic AI fails and how to mitigate those failures technically.

FINOS ensures those mitigations are governed, owned, and auditable.

NIST ensures leadership can measure and manage AI risk over time.

For organizations deploying agentic AI in production, security requires all three.

Anything less leaves a blind spot at runtime in governance, or at the board level.

1 Comment

Leave A Comment To Eve Team Cancel Comment

Name*
Message*

Scroll to top