Securiti leads GigaOm's DSPM Vendor Evaluation with top ratings across technical capabilities & business value.

View

Preparing for the Future: NIST AI RMF in the Age of Advanced AI

Contributors

Anas Baig

Product Marketing Manager at Securiti

Adeel Hasan

Sr. Data Privacy Analyst at Securiti

CIPM, CIPP/Canada

Muhammad Ismail

Assoc. Data Privacy Analyst at Securiti

Published September 12, 2024

Listen to the content

In a rapidly evolving AI landscape, the National Institute of Standards and Technology (NIST) Artificial Intelligence Risk Management Framework (AI RMF) provides a robust guideline for organizations struggling to navigate AI complexities and mitigate the risks associated with AI technologies.

The AI market is projected to reach US$184.00bn in 2024 and show an annual growth rate (CAGR 2024-2030) of 28.46%, resulting in a market volume of US$826.70bn by 2030. This figure demonstrates the sheer adoption of AI across industries and the need to assess potential risks and ethical considerations associated with AI use.

With the growing integration of AI systems across several industries, it is crucial to have a robust framework that can be utilized across geographies. The future of NIST AI RMF promises significant improvements, such as a dedicated focus on generative AI, the integration of maturity models, and alignment with international standards such as the EU AI Act and Japan AI Guidelines for Business.

These updates aim to maintain ethical, transparent, and socially responsible AI development by taking a holistic approach to AI risk management. This guide explores the projected future of NIST AI RMF and how it may affect AI governance going forward.

Deep Dive into the Future of NIST AI RMF

In response to the increasing complexity and integration of AI technologies across several industries, the NIST AI RMF is likely to undergo substantial development. These include:

Generative AI Profile

NIST has released a draft Generative AI Profile to address particular risks associated with generative AI (GenAI). This profile offers recommendations for how businesses can successfully manage these risks while considering the unique challenges that GenAI systems provide. The profile was developed with input from a large public working group and outlines 12 risk areas and several recommended actions​​. These risk areas are:

  1. Chemical, Biological, Radiological, or Nuclear (CBRN) Information or Capabilities
  2. Confabulation
  3. Dangerous, Violent, or Harmful Content
  4. Data Privacy
  5. Environmental Impacts
  6. Human-AI Configuration
  7. Information Integrity
  8. Information Security
  9. Intellectual Property
  10. Obscene, Degrading, and/or Abusive Content
  11. Harmful Bias and Homogenization
  12. Value Chain and Component Integration.

Maturity Model Integration

Integrating maturity models for AI risk management with the NIST AI RMF is becoming increasingly important. These models provide a structured approach for assessing advancement and determining areas where organizations may improve their AI governance procedures. Instead of only acting as a compliance checklist, this integration ensures that the NIST AI RMF stays applicable and useful for ongoing development.

International Alignment

NIST is integrating the AI RMF with internationally recognized guidelines and standards, such as the Japan AI Guidelines for Business and the EU AI Act, to enhance global uniformity in AI risk management. This alignment will facilitate international organizations' collaboration and ensure they can follow a unified set of best practices.

Continuous Updates and Community Engagement

The AI RMF is intended to be a dynamic document that will receive updates and be regularly reviewed in response to input from the AI community. To ensure the framework is current and useful in tackling new AI concerns, NIST has set up procedures for continuing community involvement, including workshops, public comments, and coordination with different stakeholders.

Emphasis on Trustworthiness and Responsible AI

The framework continues to prioritize AI's safe, valid, reliable, accountable, and transparent use. It encourages organizations to take a sociotechnical stance, considering both AI's social and technological implications. This strategy seeks to ensure that AI technologies are created and used transparently, without bias, and socially responsibly.

These initiatives highlight NIST's commitment to providing a robust and flexible framework that can help organizations manage AI risks in an ethical and efficient manner as the technological environment evolves. This commitment is explicitly mentioned in the NIST AI RMF document:

“The Framework and supporting resources will be updated, expanded, and improved based on evolving technology, the standards landscape around the world, and AI community experience and feedback. NIST will continue to align the AI RMF and related guidance with applicable international standards, guidelines, and practices. As the AI RMF is put into use, additional lessons will be learned to inform future updates and additional resources.”

The Role of Stakeholders in Shaping the Future of NIST AI RMF

Stakeholders are essential to the NIST AI RMF's development because they bring a variety of viewpoints and levels of experience to the table. The main methods that stakeholders have an impact on the framework are as follows:

Public and Private Sector Collaboration

The public and commercial sectors have jointly contributed to developing the NIST AI RMF. Contributions from stakeholders in academia, government, technology corporations, and civil society ensure that the framework covers the wide range of concerns related to artificial intelligence. Together, these stakeholders help create a comprehensive, well-balanced framework applicable to a wide range of industries and applications.

Public Feedback and Workshops

NIST often solicits public input via workshops, public comment periods, and requests for information. Stakeholders may use these venues to share their experiences, express their concerns, and suggest changes. The flexible process of integrating public feedback ensures that the AI RMF remains responsive to emergent challenges and stakeholder requirements.

Working Groups and Advisory Panels

Specialized working groups and advisory panels, including experts from diverse sectors, provide specific suggestions and technical assistance. These groups emphasize particular elements of AI risk management, such as GenAI or ethical issues, ensuring the framework fully covers all essential and evolving areas of concern.

International Cooperation

Stakeholders also play a significant role in aligning the AI RMF with international standards and best practices. By partnering with international agencies and adopting global viewpoints, the NIST AI RMF can be aligned with existing guidelines, easing its adoption by multinational corporations and providing consistency in AI governance across geographies.

How Securiti Can Help

Securiti’s Data Command Center enables organizations to comply with the NIST AI RMF by securing the organization’s data, enabling organizations to maximize data value, and fulfilling an organization’s obligations around data security, data privacy, data governance, and compliance.

Organizations can overcome hyperscale data environment challenges by delivering unified intelligence and controls for data across public clouds, data clouds, and SaaS, enabling organizations to swiftly comply with privacy, security, governance, and compliance requirements.

Request a demo to witness Securiti in action.

Join Our Newsletter

Get all the latest information, law updates and more delivered to your inbox



More Stories that May Interest You
Videos
View More
Mitigating OWASP Top 10 for LLM Applications 2025
Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...
View More
Top 6 DSPM Use Cases
With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...
View More
Colorado Privacy Act (CPA)
What is the Colorado Privacy Act? The CPA is a comprehensive privacy law signed on July 7, 2021. It established new standards for personal...
View More
Securiti for Copilot in SaaS
Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...
View More
Top 10 Considerations for Safely Using Unstructured Data with GenAI
A staggering 90% of an organization's data is unstructured. This data is rapidly being used to fuel GenAI applications like chatbots and AI search....
View More
Gencore AI: Building Safe, Enterprise-grade AI Systems in Minutes
As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...
View More
Navigating CPRA: Key Insights for Businesses
What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...
View More
Navigating the Shift: Transitioning to PCI DSS v4.0
What is PCI DSS? PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards to ensure safe processing, storage, and...
View More
Securing Data+AI : Playbook for Trust, Risk, and Security Management (TRiSM)
AI's growing security risks have 48% of global CISOs alarmed. Join this keynote to learn about a practical playbook for enabling AI Trust, Risk,...
AWS Startup Showcase Cybersecurity Governance With Generative AI View More
AWS Startup Showcase Cybersecurity Governance With Generative AI
Balancing Innovation and Governance with Generative AI Generative AI has the potential to disrupt all aspects of business, with powerful new capabilities. However, with...

Spotlight Talks

Spotlight 11:29
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Watch Now View
Spotlight 11:18
Rewiring Real Estate Finance — How Walker & Dunlop Is Giving Its $135B Portfolio a Data-First Refresh
Watch Now View
Spotlight 13:38
Accelerating Miracles — How Sanofi is Embedding AI to Significantly Reduce Drug Development Timelines
Sanofi Thumbnail
Watch Now View
Spotlight 10:35
There’s Been a Material Shift in the Data Center of Gravity
Watch Now View
Spotlight 14:21
AI Governance Is Much More than Technology Risk Mitigation
AI Governance Is Much More than Technology Risk Mitigation
Watch Now View
Spotlight 12:!3
You Can’t Build Pipelines, Warehouses, or AI Platforms Without Business Knowledge
Watch Now View
Spotlight 47:42
Cybersecurity – Where Leaders are Buying, Building, and Partnering
Rehan Jalil
Watch Now View
Spotlight 27:29
Building Safe AI with Databricks and Gencore
Rehan Jalil
Watch Now View
Spotlight 46:02
Building Safe Enterprise AI: A Practical Roadmap
Watch Now View
Spotlight 13:32
Ensuring Solid Governance Is Like Squeezing Jello
Watch Now View
Latest
Shrink The Blast Radius: Automate Data Minimization with DSPM View More
Shrink The Blast Radius
Recently, DaVita disclosed a ransomware incident that ultimately impacted about 2.7 million people, and it’s already booked $13.5M in related costs this quarter. Healthcare...
Why I Joined Securiti View More
Why I Joined Securiti
I’m beyond excited to join Securiti.ai as a sales leader at this pivotal moment in their journey. The decision was clear, driven by three...
View More
EU Publishes Template for Public Summaries of AI Training Content
The EU released the Explanatory Notice and Template for the Public Summary of Training Content for General-Purpose AI (GPAI) Models. Learn more.
Decoding Saudi Arabia’s Cybersecurity Risk Management Framework View More
Decoding Saudi Arabia’s Cybersecurity Risk Management Framework
Discover the Kingdom of Saudi Arabia’s National Framework for Cybersecurity Risk Management by the NCA. Learn how TLP, risk assessment and proactive strategies protect...
The Rise of AI in Financial Institutions View More
The Rise of AI in Financial Institutions: Realignment of Risk & Reward
Learn how AI is transforming financial institutions by reshaping risk management, regulatory compliance, and growth opportunities. Learn how organizations can realign risk and reward...
Redefining Data Privacy Careers in the Age of AI View More
Redefining Data Privacy Careers in the Age of AI
Securiti's whitepaper provides a detailed overview of the impact AI is poised to have on data privacy jobs and what it means for professionals...
Navigating the Minnesota Consumer Data Privacy Act (MCDPA) View More
Navigating the Minnesota Consumer Data Privacy Act (MCDPA): Key Details
Download the infographic to learn about the Minnesota Consumer Data Privacy Act (MCDPA) applicability, obligations, key features, definitions, exemptions, and penalties.
EU AI Act Mapping: A Step-by-Step Compliance Roadmap View More
EU AI Act Mapping: A Step-by-Step Compliance Roadmap
Explore the EU AI Act Mapping infographic—a step-by-step compliance roadmap to help organizations understand key requirements, assess risk, and align AI systems with EU...
The DSPM Architect’s Handbook View More
The DSPM Architect’s Handbook: Building an Enterprise-Ready Data+AI Security Program
Get certified in DSPM. Learn to architect a DSPM solution, operationalize data and AI security, apply enterprise best practices, and enable secure AI adoption...
Gencore AI and Amazon Bedrock View More
Building Enterprise-Grade AI with Gencore AI and Amazon Bedrock
Learn how to build secure enterprise AI copilots with Amazon Bedrock models, protect AI interactions with LLM Firewalls, and apply OWASP Top 10 LLM...
What's
New