Skip to main content
U.S. flag

An official website of the United States government

Official websites use .gov
A .gov website belongs to an official government organization in the United States.

Secure .gov websites use HTTPS
A lock ( ) or https:// means you’ve safely connected to the .gov website. Share sensitive information only on official, secure websites.

AI Hero Image

Artificial intelligence


Artificial Intelligence (AI) is rapidly transforming our world. Remarkable surges in AI capabilities have led to a wide range of innovations including autonomous vehicles and connected Internet of Things devices in our homes. AI is even contributing to the development of a brain-controlled robotic arm that can help a paralyzed person feel again through complex direct human-brain interfaces. These new AI-enabled systems are revolutionizing and benefitting nearly all aspects of our society and economy – everything from commerce and healthcare to transportation and cybersecurity. But the development and use of the new technologies it brings are not without technical challenges and risks. 

On October 30, President Joseph R. Biden signed an Executive Order (EO) to build U.S. capacity to evaluate and mitigate the risks of Artificial Intelligence (AI) systems to ensure safety, security, and trust, while promoting an innovative, competitive AI ecosystem that supports workers and protects consumers. Learn more about NIST's responsibilities in the EO and the creation of the U.S. Artificial Intelligence Safety Institute, including the new consortium that is being established.

NIST contributes to the research, standards and data required to realize the full promise of artificial intelligence (AI) as a tool that will enable American innovation, enhance economic security and improve our quality of life. Much of our work focuses on cultivating trust in the design, development, use and governance of artificial intelligence (AI) technologies and systems. We are doing this by:

  • Conducting fundamental research to advance trustworthy AI technologies and understand and measure their capabilities and limitations.
  • Applying AI research and innovation across NIST laboratory programs.
  • Establishing benchmarks and developing data and metrics to evaluate AI technologies.
  • Leading and participating in the development of technical AI standards.
  • Contributing to discussions and development of AI policies, including supporting the National AI Advisory Committee.
  • Hosting the NIST Trustworthy & Responsible AI Resource Center providing access to a wide range of relevant AI resources.

NIST’s AI efforts fall in several categories:

NIST’s AI portfolio includes fundamental research into and development of AI technologies — including software, hardware, architectures and human interaction and teaming — vital for AI computational trust.

AI approaches are increasingly an essential component in new research. NIST scientists and engineers use various machine learning and AI tools to gain a deeper understanding of and insight into their research. At the same time, NIST laboratory experiences with AI are leading to a better understanding of AI’s capabilities and limitations.

With a long history of devising and revising metrics, measurement tools, standards and test beds, NIST increasingly is focusing on the evaluation of technical characteristics of trustworthy AI.

NIST leads and participates in the development of technical standards, including international standards, that promote innovation and public trust in systems that use AI. A broad spectrum of standards for AI data, performance and governance are — and increasingly will be — a priority for the use and creation of trustworthy and responsible AI.

A fact sheet describes NIST's AI programs.

Cultivating Trust in AI Technologies

AI and Machine Learning (ML) is changing the way in which society addresses economic and national security challenges and opportunities. It is being used in genomics, image and video processing, materials, natural language processing, robotics, wireless spectrum monitoring and more. These technologies must be trustworthy and developed for responsible AI practice and use. Trustworthy AI systems are demonstrated to be valid and reliable, safe, secure and resilient, accountable and transparent, explainable and interpretable, privacy-enhanced, and fair with harmful bias managed.

Delivering the needed measurements, standards and other tools is a primary focus for NIST’s portfolio of AI efforts. It is an area in which NIST has special responsibilities and expertise. NIST relies heavily on stakeholder input, including via workshops, and issues most publications in draft for comment.

The Research

Projects & Programs


JARVIS-ML is a repository of machine learning (ML) model parameters, descriptors, and ML related input and target data. JARVIS-ML is a part of the NIST-JARVIS

Additional Resources Links


Minimizing Harms and Maximizing the Potential of Generative AI

As generative AI tools like ChatGPT become more commonly used, we must think carefully about the impact on people and society.

NIST Seeks Collaborators for Consortium Supporting Artificial Intelligence Safety

NIST Updates Guide on Communications Standards to Federal Agencies

Department of Commerce to Undertake Key Responsibilities in Historic Artificial Intelligence Executive Order

Bias in AI
Bias in AI
NIST contributes to the research, standards, and data required to realize the full promise of artificial intelligence (AI) as an enabler of American innovation across industry and economic sectors. Working with the AI community, NIST seeks to identify the technical requirements needed to cultivate trust that AI systems are accurate and reliable, safe and secure, explainable, and free from bias. A key but still insufficiently defined building block of trustworthiness is bias in AI-based products and systems. That bias can be purposeful or inadvertent. By hosting discussions and conducting research, NIST is helping to move us closer to agreement on understanding and measuring bias in AI systems.
Psychology of Interpretable and Explainable AI
Psychology of Interpretable and Explainable AI
 The purpose of this pre-recorded webinar, is to promote and more broadly share the release of NIST IR-8367, "Psychological Foundations of Explainability and Interpretability in Artificial Intelligence." This is a pre-recorded interview between the author of the paper, Dr. David Broniatowski, and a member of the NIST ITL team, Natasha Bansgopaul, asking key questions to highlight important insights from the paper that was published in April 2021.


Stay in Touch

Sign up for our newsletter to stay up to date with the latest research, trends, and news for Artificial intelligence.