Navigating the New Frontiers: An Overview of Evolving AI Regulations

In the rapidly advancing world of artificial intelligence (AI), the technology's potential is as vast as its risks are profound. As AI systems become increasingly sophisticated and integrated into our daily lives, governments and institutions worldwide are racing to establish robust regulatory frameworks.

This blog post explores three pivotal developments in AI regulation: the EU AI Act, the NIST AI Safety Consortium, and SR 11-7.

EU AI Act: Europe's AI Regulation

The European Union has long been at the forefront of digital regulation, and its approach to AI is no exception. The EU AI Act, proposed in April 2021, is set to become the world's first comprehensive AI law.

To recap, the EU AI Act categorizes AI systems based on their risk level:

  1. Unacceptable Risk: Systems that manipulate human behavior or enable social scoring are outright banned.
  2. High Risk: AI in critical sectors like healthcare, education, and law enforcement face stringent requirements.
  3. Limited Risk: Systems like chatbots need to be transparent about their AI nature.
  4. Minimal Risk: Most AI applications fall here, with light oversight.

The Act's risk-based approach aims to foster innovation while safeguarding fundamental rights. With substantial fines for non-compliance—up to 6% of global annual turnover—the EU is sending a clear message: AI development must prioritize human values. Learn more about the top 5 actions your company can take today to make compliance easier

NIST AI Safety Consortium: Collaborative Risk Mitigation

In the United States, the National Institute of Standards and Technology (NIST) launched the AI Safety Consortium - a public-private partnership focused on collaborative solutions rather than legislation. It brings together stakeholders to address AI safety challenges like robustness, security, interpretability, and fairness through shared best practices.

Key focus areas include:

  • Robustness: Ensuring AI systems perform reliably under varied conditions.
  • Security: Protecting AI from adversarial attacks and misuse.
  • Interpretability: Making AI decision-making processes understandable.
  • Fairness: Mitigating bias and discrimination in AI outputs.

By fostering dialogue and sharing best practices, NIST aims to cultivate a culture of responsibility in the AI industry. This voluntary, consensus-driven approach reflects the U.S.'s preference for industry self-regulation. Read the press release about the work Vectice is doing in the consortium efforts here.

SR 11-7: Foundation for AI/ML Risk Management

SR 11-7 is a robust starting point for governing AI/ML models, focusing on managing key risks. While SR 11-7 mandates detailed explanations of model methodologies and challenges used to validate the model independently, it does not mandate documentation on the feature engineering that shapes the data and significantly impacts the model results. 

Additionally, the regulation doesn’t include guidance on local and global interpretability. Both of these are under consideration for future updates to the regulation. Both of these are under consideration.

Further out, the regulation could include mandates for:

  • Data Lineage & Security: To enforce consent, privacy, protection, and security of personal data, mandates may evolve to include strong data controls, policies, and governance around collection, lineage, and quality. 
  • Consumer Protection: AI is powerful at detecting patterns but can be misled when the underlying data contains inherent biases. These biases could inadvertently lead to discrimination, improper personalization, and exclusion from certain products. The future will likely include regulations to protect consumers from these mishaps. (Source: KPMG)

Final Thoughts

As we navigate these new frontiers in AI regulation, it is clear that a balanced approach is crucial. The EU AI Act exemplifies a stringent, risk-based regulatory framework, ensuring innovation does not come at the cost of human rights. In contrast, the NIST AI Safety Consortium emphasizes collaborative risk mitigation and industry self-regulation, fostering a culture of responsibility without stifling innovation. Meanwhile, SR 11-7 provides a foundational guideline for managing AI/ML risks, with future updates poised to enhance its robustness and inclusivity.

The convergence of these regulatory approaches underscores the global commitment to harnessing AI's potential while mitigating its risks. For businesses, this evolving landscape presents both challenges and opportunities. Compliance with these regulations not only safeguards against legal repercussions but also builds consumer trust and promotes sustainable innovation.

As AI continues to permeate various sectors, staying informed and proactive about these regulatory changes is essential. By aligning AI development with these emerging standards, organizations can contribute to a safer, more ethical, and innovative future. The journey ahead may be complex, but with a clear focus on responsible AI practices, we can unlock the transformative power of AI while ensuring its benefits are broadly shared and its risks effectively managed.

Back to Blog