About 553,000 results
Open links in new tab
  1. Agentic AI Security: Threats, Defenses, Evaluation, and Open

    In this survey, we focus on this very timely and pertinent problem of agentic AI security, and present the current state-of-the-art in novel AI agent attack methodologies, defense strategies, …

  2. [2502.19328] Agentic Reward Modeling: Integrating Human …

    Feb 26, 2025 · In this paper, we propose agentic reward modeling, a reward system that combines reward models with verifiable correctness signals from different aspects to provide …

  3. Jul 9, 2025 · Despite growing awareness of this problem, systematic detection and mitigation approaches remain limited. This paper presents a large-scale empirical study of reward …

  4. Beyond Accuracy: A Multi-Dimensional Framework for Evaluating ...

    Abstract Current agentic AI benchmarks predominantly evaluate task completion accuracy, while overlooking critical enterprise requirements such as cost-efficiency, reliability, and operational …

  5. Adaptive Monitoring and Real‑World Evaluation of Agentic AI

    Abstract Agentic artificial intelligence (AI) — multi‑agent systems that combine large language models with external tools and autonomous planning — are rapidly transitioning from research …

  6. The Real Barrier to LLM Agent Usability is Agentic ROI - arXiv.org

    May 23, 2025 · We outline the roadmap across different development stages to bridge the current usability gaps, aiming to make LLM agents truly scalable, accessible, and effective in real …

  7. GitHub - junhua/awesome-llm-agents: A Collection of High …

    The detailed thought process of forming this project is documented at this Medium Post. It's put behind a paywall to prevent the evil LLMs' crawling. The full category breakdown. Retrieval …

  8. Reward models (RMs) are crucial for the train- ing and inference-time scaling up of large lan- guage models (LLMs). However, existing re- ward models primarily focus .

  9. GitHub - THU-KEG/Agentic-Reward-Modeling: [ACL 2025] Agentic Reward

    We empirically implement a reward agent in this repo, named RewardAgent, that combines human preference rewards with two verifiable signals: factuality and instruction following, to …

  10. to generate and perpetually evolve a comprehensive safety bench-mark. SafeEvalAgent leverages a synergistic pipeline of specialized agents and incorporates a Self-evolving …