Today concerted attention across the AI ecosystem focuses on creating effective AI agents. Increasingly capable AI agents promise great opportunities for economic competitiveness, but also require their developers, deployers, and users to manage security and reliability risks. AI agents can perceive and take actions in environments; the leading AI agent paradigm today embeds general-purpose AI models into systems with software scaffolding that enable a model to manipulate tools to take actions beyond simple text output. AI agents are increasingly deployed as experimental products that can build software applications, browse the internet, and more.
To date there has been no attempt to provide a comprehensive taxonomy of these agent tools. Such a taxonomy could enable actors across the AI supply chain to more clearly share information about system capabilities and considerations. For example, this can enable an AI agent developer to share tool capabilities and limitations with downstream developers to create applications that make full use of agent capabilities. It can similarly support third-party researchers and users to report flaws or incidents with categories of AI agent tools. A shared vocabulary can support this communication.
To take steps toward providing this resource, CAISI and NIST hosted an AISIC workshop with approximately 140 experts in January. Below we present lessons learned from the community through that workshop.
Participants identified various approaches to structure a taxonomy of tool use, including:
Each approach above has its strengths and weaknesses. Some approaches may invoke things beyond the tool. A risk-based taxonomy, for example, will depend on deployment conditions; access patterns, autonomy, and other approaches may depend on the ways in which a tool is implemented in practice. Rather than homing in on one single taxonomy, workshop participants raised these multiple approaches, and some suggested that multidimensional intersections across multiple taxonomies would be a promising direction for future work. In practice, taxonomies may build upon or otherwise complement each other: One structured around monitoring may be informed by a risk-based taxonomy, which in turn may be informed by functionality-focused or access patterns. Stakeholders may benefit from creating taxonomies of tool use to fit their particular needs.
Consider below two taxonomies that address approaches identified in the AISIC workshop. The first takes a functional approach, categorizing tools by what they enable the model to do. Reflecting workshop discussions, it aims for comprehensive coverage for the categories of tools that enable types of actions. The taxonomy provides a clear baseline that can be expanded or tailored to particular needs. Developers may use these types as a structured way to reason about the capabilities of their agent systems during system development, and to subsequently communicate externally on what types of actions are possible and which ones may be constrained. Indeed, within each category, example tools may be more or less constrained, e.g., sensors may be subject to filters that reduce the risk of indirect prompt injections from search results.
Figure 1. Functionality-Oriented Taxonomy of Tools in AI Agent Systems
Purpose | Type | Examples |
Perception: ways a model may perceive the environment | Sensors | Internal database, monitoring, diagnostics, GUI, voice, internet search, physical world |
Reasoning: ways that a model may reason beyond inference | Planning | Task-decomposition, path-finding models |
Analysis | Scratchpads, calculators, simulations | |
Resource management | Memory, self-management | |
Action: ways that a model may directly affect the environment | Authentication | Login, CAPTCHA, wallet |
Computer use | Application-specific GUI interaction, website interactions, computer use | |
Running code | Sandboxed code interpreter, IDE, file operations, code execution | |
Software extensions | Calendar, social media API | |
Physical extensions | Robotic arm, laboratory tools in factory setting, robot in an open environment | |
Human interaction | Phone calls | |
Agent interaction | Multi-agent simulation, sub-agents that can interact with outside world, third-party agent interactions |
The second taxonomy addresses the constraints that may limit the actions possible with tool use. It considers constraints as a function of tool permissions and the action environment. Some tools may enable read-only actions, while others enable (“write”) actions that impact state. In practice, many agent implementations may limit write access by using tools with restricted interactions or constraining otherwise plausibly unlimited tools like code execution. Some agent implementations may access untrusted resources like the open internet, whereas others are designed for deployment in sanitized settings. This taxonomy reflects access patterns raised in the workshop. Stakeholders may deepen these categories with additional gradations of “write” permissions or trust levels. Such efforts, together with specific knowledge of deployments, may be useful to develop a specific agent risk taxonomy or to perform a risk assessment, complementing additional resources like NIST AI 600-1.
Figure 2. Taxonomy of Constrained Tool Access Patterns with Example AI Agent Systems
Tool Permissions/ Environment | Read Only | (Constrained) Write | Write |
Trusted Environments | RAG | Application-specific GUI or API use | Coding agent in a trusted repository |
Untrusted Environments | Deep research | Browser use | Computer use |
The Consortium contributed valuable expertise to shape these taxonomies and identify additional approaches that can be expanded upon in the future. We invite your feedback and encourage you to adapt and expand upon these findings to serve your needs. Tool taxonomies are one method to improve transparency on capabilities and deployments along the AI agent value chain. We welcome your engagement as we evaluate how best to support stakeholders in AI agent development and deployment. You can share comments via email to CAISI-agents [at] nist.gov.