Skip to main content
U.S. flag

An official website of the United States government

Official websites use .gov
A .gov website belongs to an official government organization in the United States.

Secure .gov websites use HTTPS
A lock ( ) or https:// means you’ve safely connected to the .gov website. Share sensitive information only on official, secure websites.

NIST Information Technology Laboratory AI Webinar Series: Building Measurement Probes into Agentic AI Ecosystems

AI agents represent the next evolution of language models, capable of planning multi-step tasks, utilizing tools, searching databases, and making decisions autonomously. As organizations integrate these agents into higher-stakes ecosystems, there is an urgent need for measurement tools that can trace the decisions agentic models make and verify their appropriateness. Traceability depends upon the ability to reconstruct the chain of reasoning, usage of tools, and other evidence behind an agent's decision.

The NIST Information Technology Laboratory (ITL) AI Program will host a technical webinar on early research focused on developing automated measurement tools, called probes, to build traceability into agentic AI ecosystems. The approach adapts established techniques such as judges/verifiers, grounds them in a knowledge-base, and empowers them to evaluate agentic AI output.

The webinar will discuss existing technical gaps in the measurement infrastructure of agentic AI systems. The team will outline initial research into a promising approach for applying concepts such as adversarial verifiers to better evaluate Agentic AI outputs.

During and following the webinar, participants will be encouraged to identify practical measurement challenges they face with agentic AI. The ITL team is seeking suggestions for potential use cases – including but not limited to those which are domain-specific – that would benefit from dynamic verification and help refine and expand the development of automated measurement of agentic AI systems.

About the ITL AI Program

The AI Program in NIST’s Information Technology Laboratory (ITL) accelerates and expands development and adoption of AI by strengthening trust in AI through vital measurement science, testing and evaluation, guidance, and standards.

Focus areas include:

  • Advancing Testing, Evaluation, Verification, and Validation (TEVV) to ensure that AI is deployed and used responsibly, reliably, and efficiently
  • Providing resources for managing AI benefits and risks, empowering industry, academia, non-profits, and government to make informed decisions about AI trustworthiness and use
  • Positioning U.S. as preeminent in AI technical and governance standards, ensuring the U.S. leads global AI innovation
  • Enabling U.S. to lead in applying AI to high-priority areas – including manufacturing and cybersecurity for critical infrastructure – via innovative approaches to address measurement challenges
Created March 12, 2026, Updated March 16, 2026
Was this page helpful?