NIST's Latest Guidance on Secure AI Development and Global Standards

Baker Botts L.L.P.
Contact

Baker Botts L.L.P.

On July 26, the National Institute of Standards and Technology (NIST) released four guidance documents related to artificial intelligence (AI) development and implementation. These documents were issued pursuant to instructions under AI Executive Order 14110, which directed several U.S. government agencies to promulgate guidance and regulations with respect to safe, secure, and trustworthy AI.

The first document, titled "Artificial Intelligence Risk Management Framework: Generative Artificial Intelligence Profile" (RMF GAI), describes and defines risks associated with generative AI (GAI) and outlines how organizations can govern, manage, and mitigate such risks. The RMF GAI profiles the functions and categories of NIST's 2023 AI Risk Management Framework, specifically for GAI technology. It provides a cross-sectoral profile for managing risks related to GAI implementation, applicable across different sectors and addressing both current concerns and potential future harmful scenarios.

NIST also released three additional documents:

"Secure Software Development Practices for Generative AI and Dual-Use Foundation Models" (SSDF), which updates prior NIST software development guidance to add recommendations for implementing secure development practices specifically tailored to generative AI systems. It covers the entire AI model development lifecycle and emphasizes practices to secure AI elements and mitigate risks from malicious tampering.

"A Plan for Global Engagement on AI Standards" (AI Plan), which provides directives to drive worldwide development and implementation of AI-related consensus standards, cooperation, and information sharing. It emphasizes the need for context-sensitive, performance-based, and human-centered AI standards.

"Managing Misuse Risk for Dual-Use Foundation Models" (MMRD), which offers comprehensive guidelines for identifying, measuring, and mitigating misuse risks associated with powerful AI models. This document is open for public comment through September 9, 2024.

Together, these guidance documents attempt to define best practices to reduce risks that arise when developing and deploying AI models. While the NIST guidance is not legally binding, those developing or deploying AI models might take note, as deviation from prevailing practices or recommendations could introduce insurance or liability risks, particularly for those operating in accordance with federal information systems.

This document defines risks that are novel to or exacerbated by the use of GAI. After introducing and describing these risks, the document provides a set of suggested actions to help organizations govern, map, measure, and manage these risks.

nvlpubs.nist.gov/...

 

DISCLAIMER: Because of the generality of this update, the information provided herein may not be applicable in all situations and should not be acted upon without specific legal advice based on particular situations. Attorney Advertising.

© Baker Botts L.L.P.

Written by:

Baker Botts L.L.P.
Contact
more
less

PUBLISH YOUR CONTENT ON JD SUPRA NOW

  • Increased visibility
  • Actionable analytics
  • Ongoing guidance

Baker Botts L.L.P. on:

Reporters on Deadline

"My best business intelligence, in one easy email…"

Your first step to building a free, personalized, morning email brief covering pertinent authors and topics on JD Supra:
*By using the service, you signify your acceptance of JD Supra's Privacy Policy.
Custom Email Digest
- hide
- hide