Nov 2 -- The National Institute of Standards and Technology (NIST), an agency of the United States Department of Commerce, in support of efforts to create safe and trustworthy artificial intelligence (AI), is establishing the Artificial Intelligence Safety Institute Consortium (“Consortium”). The Consortium will help equip and empower the collaborative establishment of a new measurement science that will enable the identification of proven, scalable, and interoperable techniques and metrics to promote development and responsible use of safe and trustworthy AI, particularly for the most advanced AI systems, such as the most capable foundation models.
This notice is the initial step for NIST in collaborating with non-profit organizations, universities, other government agencies, and technology companies to address challenges associated with the development and deployment of AI. Many of these challenges were identified under the Executive Order of October 30, 2023 (The Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence) and the NIST AI RMF Roadmap. Much of this research will center on evaluations of and approaches towards safer, more trustworthy AI systems.
NIST invites organizations to provide letters of interest describing technical expertise and products, data, and/or models to support and demonstrate pathways to enable safe and trustworthy artificial intelligence (AI) systems through the AI Risk Management Framework (AI RMF). This is the initial step for NIST in collaborating with technology companies, other government agencies, and non-profit organizations to address challenges identified under the NIST AI RMF roadmap regarding evaluation of and approaches towards safer AI systems.
Participation in the consortium is open to all interested organizations that can contribute their expertise, products, data, and/or models to the activities of the consortium. Selected participants will be required to enter into a consortium Cooperative Research and Development Agreement (CRADA) with NIST. At NIST's discretion, entities which are not permitted to enter into CRADAs pursuant to law may be allowed to participate in the Consortium pursuant to separate non-CRADA agreement.
Interested organizations with relevant technical capabilities should complete a Letter of Interest (LOI) by December 2, 2023. Each responding organization’s letter of interest should include the address, point of contact, and following information:
-- The role(s) the organization will play in the consortium efforts.
-- The specific expertise will they intend to bring to the consortium.
-- The products, services, data, or other technical capabilities will they use in consortium activities.
NIST will host a workshop on November 17, 2023, to engage in a conversation about artificial intelligence (AI) safety. The hybrid workshop will be held at the Department of Commerce in Washington, DC, with options for virtual or in person attendance.
NIST supports the United States in developing standards around emerging technologies, including artificial intelligence and related systems. The NIST AI Risk Management Framework (AI RMF) provides a foundational set of approaches for holistically assessing risk for the use of AI systems. However, in deploying this framework, specific improvements in our ability to evaluate and validate AI systems are necessary, as detailed in the AI RMF roadmap, available at
https://www.nist.gov/itl/ai-risk-management-framework/roadmap-nist-artificial-intelligence-risk-management-framework-ai.
In addition, The Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence Executive Order articulated a broad set of priorities relevant to and directive of NIST's work related to AI safety and trust. NIST seeks to leverage the additional resources and capabilities made available through this consortium to meet the requirements of the Executive Order and fulfill those priorities in the future.
NIST is soliciting responses from all sources of relevant technical capabilities (see below) to enter into a consortium Cooperative Research and Development Agreement (CRADA) to provide technical expertise and products, data, and/or models to enable safe and trustworthy artificial intelligence (AI) systems. The Consortium will help enable the identification of proven, scalable, and interoperable techniques and metrics to promote development of trustworthy AI and its responsible use. The full project can be viewed at:
https://www.nist.gov/artificial-intelligence/artificial-intelligence-safety-institute.
The project is in support of the AI RMF roadmap and The Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence Executive Order.
Artificial Intelligence (AI) tools and applications are growing at an unprecedented pace, changing our way of life, and having significant impacts on society and all sectors of the economy. Yet, the potential technical and societal benefits and risks of AI require much closer examination and a more complete understanding. Aligning AI with our societal norms and values and keeping the public safe requires a broad human-centered focus, specific policies, processes, and guardrails informed by community stakeholders across various levels of our society, and bold commitment from the public sector.
To manage the broad risks of AI technologies, help to protect the public and our planet, reduce market uncertainties, and encourage even more extraordinary AI technological innovations, the National Institute of Standards and Technology (NIST) is expanding its AI measurement efforts by harnessing the broader community's interests and capabilities. NIST aims to help enable the identification of proven, scalable, and interoperable measurements and methodologies to promote development of trustworthy AI and its responsible use. This is a critical challenge at a pivotal time—not only for AI technologists but for society.
Building upon its long track record of working with the private and public sectors and its history of reliable and practical measurement and standards-oriented solutions, NIST seeks research collaborators who can support this vital undertaking. Specifically, NIST looks to
-- Create a convening space for collaborators to have an informed dialogue and enable sharing of information and knowledge
-- Engage in collaborative research and development through shared projects
-- Enable assessment and evaluation of test systems and prototypes to inform future AI measurement efforts
To create a lasting approach for continued joint research and development, NIST will engage stakeholders via this consortium. The work of the consortium will be open and transparent and provide a hub for interested parties to work together in building and maturing a measurement science for Trustworthy and Responsible AI. Consortium members will be expected to contribute technical expertise in one or more of the following areas:
-- Data and data documentation
-- AI Metrology
-- AI Governance
-- AI Safety
-- Trustworthy AI
-- Responsible AI
-- AI system design and development
-- AI system deployment
-- AI Red Teaming
-- Human-AI Teaming and Interaction
-- Test, Evaluation, Validation and Verification methodologies
-- Socio-technical methodologies
-- AI Fairness
-- AI Explainability and Interpretability
-- Workforce skills
-- Psychometrics
-- Economic analysis
-- Models, data and/or products to support and demonstrate pathways to enable safe and trustworthy artificial intelligence (AI) systems through the AI risk management framework
-- Infrastructure support for consortium projects
-- Facility space and handling of hosting consortium researchers, workshops and conferences
This project is in service of the priorities and taskings defined in The Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence Executive Order. Additionally, some of the outcomes of this project will be in support of research and development to advance the AI RMF roadmap. The consortium will be responsible for an array of efforts related to safe and trustworthy AI, including to:
1. Develop new guidelines, tools, methods, protocols and best practices to facilitate the evolution of industry standards for developing or deploying AI in safe, secure, and trustworthy ways
2. Develop guidance and benchmarks for identifying and evaluating AI capabilities, with a focus on capabilities that could potentially cause harm
3. Develop approaches to incorporate secure-development practices for generative AI, including special considerations for dual-use foundation models, including:
a. guidance related to assessing and managing the safety, security, and trustworthiness of models and related to privacy-preserving machine learning;
b. guidance to ensure the availability of testing environments
4. Develop and ensure the availability of testing environments
5. Develop guidance, methods, skills and practices for successful red-teaming and privacy-preserving machine learning
6. Develop guidance and tools for authenticating digital content
7. Develop guidance and criteria for AI workforce skills, including risk identification and management, test, evaluation, validation, and verification (TEVV), and domain-specific expertise
8. Explore the complexities at the intersection of society and technology, including the science of how humans make sense of and engage with AI in different contexts
9. Develop guidance for understanding and managing the interdependencies between and among AI actors along the lifecycle
NIST AI Safety Institute
https://www.nist.gov/artificial-intelligence/artificial-intelligence-safety-institute
NIST Press release:
https://www.nist.gov/news-events/news/2023/11/nist-seeks-collaborators-consortium-supporting-artificial-intelligence
FRN:
https://www.federalregister.gov/d/2023-24216