PulseAugur
LIVE 09:51:48
research · [30 sources] ·
0
research

Researchers propose new frameworks for securing AI agents and multi-agent systems

Multiple research papers released in April 2026 address the growing security challenges in autonomous AI agent systems. These papers propose frameworks and methodologies for enhancing the safety, trustworthiness, and governance of interacting AI agents, particularly in high-stakes domains like cybersecurity and enterprise systems. Key themes include decentralized architectures, formal verification methods, runtime safety enforcement, and robust auditing mechanisms to mitigate risks such as adversarial attacks, data poisoning, and unauthorized actions. AI

Summary written by gemini-2.5-flash-lite from 30 sources. How we write summaries →

IMPACT These frameworks aim to improve the security and trustworthiness of AI agents, potentially accelerating their adoption in critical applications.

RANK_REASON Multiple research papers published on arXiv proposing new frameworks and methodologies for AI agent security.

Read on arXiv cs.AI →

COVERAGE [30]

  1. arXiv cs.AI TIER_1 · Majed El Helou, Benjamin Ryder, Chiara Troiani, Jean Diaconu, Herv\'e Muyal, Marcelo Yannuzzi ·

    Hybrid Inspection and Task-Based Access Control in Zero-Trust Agentic AI

    arXiv:2605.02682v1 Announce Type: new Abstract: Authorizing Large Language Model (LLM)-driven agents to dynamically invoke tools and access protected resources introduces significant security risks, and the risks grow dramatically as agents engage in multi-turn conversations and …

  2. arXiv cs.AI TIER_1 · Marcelo Yannuzzi ·

    Hybrid Inspection and Task-Based Access Control in Zero-Trust Agentic AI

    Authorizing Large Language Model (LLM)-driven agents to dynamically invoke tools and access protected resources introduces significant security risks, and the risks grow dramatically as agents engage in multi-turn conversations and scale toward distributed collaboration. A compro…

  3. arXiv cs.AI TIER_1 · Luyao Xu, Xiang Chen ·

    Security Attack and Defense Strategies for Autonomous Agent Frameworks: A Layered Review with OpenClaw as a Case Study

    arXiv:2604.27464v1 Announce Type: cross Abstract: Autonomous agent frameworks built upon large language models (LLMs) are evolving into complex, tool-integrated, and continuously operating systems, introducing security risks beyond traditional prompt-level vulnerabilities. As thi…

  4. arXiv cs.AI TIER_1 · Akshay Mittal, Elyson De La Cruz ·

    Agent Name Service (ANS): A Proof-of-Concept Trust Layer for Secure AI Agent Discovery, Identity, and Governance in Kubernetes

    arXiv:2604.26997v1 Announce Type: cross Abstract: Autonomous AI agent ecosystems require stronger mechanisms for secure discovery, identity verification, capability attestation, and policy governance. Current deployments frequently lack (1) uniform agent discovery, (2) cryptograp…

  5. arXiv cs.AI TIER_1 · Yuan Sun ·

    Safe Bilevel Delegation (SBD): A Formal Framework for Runtime Delegation Safety in Multi-Agent Systems

    arXiv:2604.27358v1 Announce Type: new Abstract: As large language model (LLM) agents are deployed in high-stakes environments, the question of how safely to delegate subtasks to specialized sub-agents becomes critical. Existing work addresses multi-agent architecture selection at…

  6. arXiv cs.AI TIER_1 · Andrei Kojukhov, Arkady Bovshover ·

    Agentic AI for Cybersecurity: A Meta-Cognitive Architecture for Governable Autonomy

    arXiv:2602.11897v3 Announce Type: replace-cross Abstract: Cybersecurity decision-making increasingly occurs in environments characterized by uncertainty, partial observability, and adversarial manipulation, where heterogeneous signals from multiple sources are often incomplete, a…

  7. arXiv cs.AI TIER_1 · Yu-Chao Huang, Zhen Tan, Mohan Zhang, Pingzhi Li, Zhuo Zhang, Tianlong Chen ·

    TRUST: A Framework for Decentralized AI Service v.0.1

    arXiv:2604.27132v1 Announce Type: new Abstract: Large Reasoning Models (LRMs) and Multi-Agent Systems (MAS) in high-stakes domains demand reliable verification, yet centralized approaches suffer four limitations: (1) Robustness, with single points of failure vulnerable to attacks…

  8. arXiv cs.AI TIER_1 · Christian Schroeder de Witt, Klaudia Krawiecka, Igor Krawczuk, Ben Hagag, William L. Anderson, Peter Belcak, Ben Bucknall, Xiaohong Cai, Ayush Chopra, Doron Cohen, Ron F. Del Rosario, Andis Draguns, Annie Gray, Keren Katz, Vasilios Mavroudis, Jaron Mink, ·

    Open Challenges in Multi-Agent Security: Towards Secure Systems of Interacting AI Agents

    arXiv:2505.02077v2 Announce Type: replace-cross Abstract: AI agents are beginning to interact with each other directly and across internet platforms and physical environments, creating security challenges beyond traditional cybersecurity and AI safety frameworks. Free-form protoc…

  9. arXiv cs.AI TIER_1 · Hung Dang ·

    Enforcing Benign Trajectories: A Behavioral Firewall for Structured-Workflow AI Agents

    arXiv:2604.26274v1 Announce Type: cross Abstract: Structured-workflow agents driven by large language models execute tool calls against sensitive external environments. We propose \codename, a telemetry-driven behavioral anomaly detection firewall. Drawing on sequence-based intru…

  10. arXiv cs.AI TIER_1 · Hung Dang ·

    Enforcing Benign Trajectories: A Behavioral Firewall for Structured-Workflow AI Agents

    Structured-workflow agents driven by large language models execute tool calls against sensitive external environments. We propose \codename, a telemetry-driven behavioral anomaly detection firewall. Drawing on sequence-based intrusion detection, \codename\ compiles verified benig…

  11. arXiv cs.AI TIER_1 · Ignacio Peyrano ·

    From CRUD to Autonomous Agents: Formal Validation and Zero-Trust Security for Semantic Gateways in AI-Native Enterprise Systems

    Enterprise software engineering is shifting away from deterministic CRUD/REST architectures toward AI-native systems where large language models act as cognitive orchestrators. This transition introduces a critical security tension: probabilistic LLMs weaken classical mechanisms …

  12. arXiv cs.AI TIER_1 · Yixiang Zhang, Xinhao Deng, Jiaqing Wu, Yue Xiao, Ke Xu, Qi Li ·

    AgentWard: A Lifecycle Security Architecture for Autonomous AI Agents

    arXiv:2604.24657v1 Announce Type: cross Abstract: Autonomous AI agents extend large language models into full runtime systems that load skills, ingest external content, maintain memory, plan multi-step actions, and invoke privileged tools. In such systems, security failures rarel…

  13. arXiv cs.CL TIER_1 · Mikko Lempinen, Joni Kemppainen, Niklas Raesalmi ·

    AVISE: Framework for Evaluating the Security of AI Systems

    arXiv:2604.20833v2 Announce Type: replace-cross Abstract: As artificial intelligence (AI) systems are increasingly deployed across critical domains, their security vulnerabilities pose growing risks of high-profile exploits and consequential system failures. Yet systematic approa…

  14. arXiv cs.AI TIER_1 · Tam Nguyen, Moses Ndebugre, Dheeraj Arremsetty ·

    Security Considerations for Multi-agent Systems

    arXiv:2603.09002v2 Announce Type: replace-cross Abstract: Multi-agent artificial intelligence systems or MAS are systems of autonomous agents that exercise delegated tool authority, share persistent memory, and coordinate via inter-agent communication. MAS introduces qualitativel…

  15. arXiv cs.AI TIER_1 · Jiaqi Li, Yang Zhao, Bin Sun, Yang Yu, Jian Chang, Lidong Zhai ·

    Poster: ClawdGo: Endogenous Security Awareness Training for Autonomous AI Agents

    arXiv:2604.24020v1 Announce Type: cross Abstract: Autonomous AI agents deployed on platforms such as OpenClaw face prompt injection, memory poisoning, supply-chain attacks, and social engineering, yet existing defences address only the platform perimeter, leaving the agent's own …

  16. arXiv cs.LG TIER_1 · Ben Hagag, William L. Anderson, Christian Schroeder de Witt, Sarah Scheffler ·

    Architecture Matters for Multi-Agent Security

    arXiv:2604.23459v1 Announce Type: cross Abstract: Multi-agent systems (MAS), composed of networks of two or more autonomous AI agents, have become increasingly popular in production deployments, yet introduce security risks that do not arise in single-agent settings. Even if indi…

  17. arXiv cs.LG TIER_1 · Kexin Chu ·

    From Stateless Queries to Autonomous Actions: A Layered Security Framework for Agentic AI Systems

    arXiv:2604.23338v1 Announce Type: cross Abstract: Agentic AI systems face security challenges that stateless large language models do not. They plan across extended horizons, maintain persistent memory, invoke external tools, and coordinate with peer agents. Existing security ana…

  18. Hugging Face Daily Papers TIER_1 ·

    Structured Security Auditing and Robustness Enhancement for Untrusted Agent Skills

    Agent Skills package SKILL.md files, scripts, reference documents, and repository context into reusable capability units, turning pre-load auditing from single-prompt filtering into cross-file security review. Existing guardrails often flag risk but recover malicious intent incon…

  19. arXiv cs.AI TIER_1 · Qi Li ·

    AgentWard: A Lifecycle Security Architecture for Autonomous AI Agents

    Autonomous AI agents extend large language models into full runtime systems that load skills, ingest external content, maintain memory, plan multi-step actions, and invoke privileged tools. In such systems, security failures rarely remain confined to a single interface; instead, …

  20. Hugging Face Daily Papers TIER_1 ·

    A Comparative Evaluation of AI Agent Security Guardrails

    This report presents a comparative evaluation of DKnownAI Guard in AI agent security scenarios, benchmarked against three competing products: AWS Bedrock Guardrails, Azure Content Safety, and Lakera Guard. Using human annotation as the ground truth, we assess each guardrail's abi…

  21. Hugging Face Daily Papers TIER_1 ·

    Poster: ClawdGo: Endogenous Security Awareness Training for Autonomous AI Agents

    Autonomous AI agents deployed on platforms such as OpenClaw face prompt injection, memory poisoning, supply-chain attacks, and social engineering, yet existing defences address only the platform perimeter, leaving the agent's own threat judgement entirely untrained. We present Cl…

  22. Hugging Face Daily Papers TIER_1 ·

    Structural Enforcement of Goal Integrity in AI Agents via Separation-of-Powers Architecture

    Recent evidence suggests that frontier AI systems can exhibit agentic misalignment, generating and executing harmful actions derived from internally constructed goals, even without explicit user requests. Existing mitigation methods, such as Reinforcement Learning from Human Feed…

  23. Mastodon — sigmoid.social TIER_1 · [email protected] ·

    Joint guidance just released from leading Western security agencies on safely adopting agentic AI services. Key considerations, emerging risks, and best practic

    Joint guidance just released from leading Western security agencies on safely adopting agentic AI services. Key considerations, emerging risks, and best practices for implementation. Worth a read. #AI #AgenticAI #Cybersecurity #ASD #ACSC #CISA #NSA #CCCS #NCSCNZ #NCSCUK Careful A…

  24. Mastodon — sigmoid.social TIER_1 Italiano(IT) · [email protected] ·

    AI runtime security for the agentic enterprise # AgenticAI # AgenticArtificialIntelligence # AI # ArtificialIntelligence #

    https://www. europesays.com/2954859/ AI runtime security for the agentic enterprise # AgenticAI # AgenticArtificialIntelligence # AI # ArtificialIntelligence # EnterprisesTurnToRuntimeSecurityToCloseTheAgenticAITrustGap # KellyKnight # SiliconANGLE

  25. Mastodon — sigmoid.social TIER_1 · [email protected] ·

    https://www. europesays.com/2954857/ SAS cements agentic AI-ready data management foundation # AgenticAI # AgenticArtificialIntelligence # AI # ArtificialIntell

    https://www. europesays.com/2954857/ SAS cements agentic AI-ready data management foundation # AgenticAI # AgenticArtificialIntelligence # AI # ArtificialIntelligence

  26. Mastodon — fosstodon.org TIER_1 · [email protected] ·

    CISA, NSA, and partner agencies published fresh guidance on secure AI agent adoption. We analyzed what it means for identity controls, privilege boundaries, and

    CISA, NSA, and partner agencies published fresh guidance on secure AI agent adoption. We analyzed what it means for identity controls, privilege boundaries, and deployment governance in real systems. https:// go.aintelligencehub.com/ma-usc yberagenciesaiagen # AI # Cybersecurity …

  27. Mastodon — fosstodon.org TIER_1 · [email protected] ·

    Joint guidance just released from leading Western security agencies on safely adopting agentic AI services. Key considerations, emerging risks, and best practic

    Joint guidance just released from leading Western security agencies on safely adopting agentic AI services. Key considerations, emerging risks, and best practices for implementation. Worth a read. #AI #AgenticAI #Cybersecurity #ASD #ACSC #CISA #NSA #CCCS #NCSCNZ #NCSCUK Careful A…

  28. Mastodon — mastodon.social TIER_1 · [email protected] ·

    Agentic AI security alert ⚠️ CISA guidance: Autonomous AI systems expand attack surface beyond traditional controls. Risks: • Unchecked actions • API exposure •

    Agentic AI security alert ⚠️ CISA guidance: Autonomous AI systems expand attack surface beyond traditional controls. Risks: • Unchecked actions • API exposure • Data leakage AI = operational entity now. Source: https://www. cisa.gov/resources-tools/resou rces/careful-adoption-age…

  29. Mastodon — mastodon.social TIER_1 Français(FR) · [email protected] ·

    Autonomous AI agents write code, call APIs, orchestrate systems... often without human supervision. The question of trust in this co

    Les agents IA autonomes écrivent du code, appellent des APIs, orchestrent des systèmes… souvent sans supervision humaine. La question de confiance dans cette couche "agentique" devient centrale en sécurité. Qui audite ce que l'agent a décidé tout seul à 3h du matin ? 🤖 La surface…

  30. Mastodon — mastodon.social TIER_1 · [email protected] ·

    Megent Megent: Open-Source Runtime Policy for AI Agents. Megent helps secure AI agents by enforcing allow/deny tool policies, masking sensitive data, and loggin

    Megent Megent: Open-Source Runtime Policy for AI Agents. Megent helps secure AI agents by enforcing allow/deny tool policies, masking sensitive data, and logging every tool decision. A zero-trust safety layer for AI agents. Wrap tools fast. Ship with confidence. Control every age…