{"id":16183,"date":"2026-05-07T02:51:22","date_gmt":"2026-05-07T02:51:22","guid":{"rendered":"https:\/\/newestek.com\/?p=16183"},"modified":"2026-05-07T02:51:22","modified_gmt":"2026-05-07T02:51:22","slug":"us-government-agency-to-safety-test-frontier-ai-models-before-release","status":"publish","type":"post","link":"https:\/\/newestek.com\/?p=16183","title":{"rendered":"US government agency to safety test frontier AI models before release"},"content":{"rendered":"<div>\n<div id=\"remove_no_follow\">\n<div class=\"grid grid--cols-10@md grid--cols-8@lg article-column\">\n<div class=\"col-12 col-10@md col-6@lg col-start-3@lg\">\n<div class=\"article-column__content\">\n<section class=\"wp-block-bigbite-multi-title\">\n<div class=\"container\"><\/div>\n<\/section>\n<p>The Center for AI Standards and Innovation (CAISI), a division of the US Department of Commerce, has signed agreements with Google DeepMind, Microsoft, and xAI that would give the agency the ability to vet AI models from these organizations and others prior to their being made publicly available.<\/p>\n<p>According to a <a href=\"https:\/\/www.nist.gov\/news-events\/news\/2026\/05\/caisi-signs-agreements-regarding-frontier-ai-national-security-testing\" target=\"_blank\" rel=\"noreferrer noopener\">release<\/a> from CAISI, which is part of the department\u2019s National Institute of Standards and Technology (<a href=\"https:\/\/www.csoonline.com\/article\/4134743\/us-dominance-of-agentic-ai-at-the-heart-of-new-nist-initiative.html\" target=\"_blank\">NIST<\/a>), it will \u201cconduct pre-deployment evaluations and targeted research to better assess frontier AI capabilities and advance the state of AI security.\u201d<\/p>\n<p>The three join Anthropic and OpenAI, which signed <a href=\"https:\/\/www.nist.gov\/news-events\/news\/2024\/08\/us-ai-safety-institute-signs-agreements-regarding-ai-safety-research\" target=\"_blank\" rel=\"noreferrer noopener\">similar agreements<\/a> almost two years ago during the Biden administration, when CAISI was known as the US Artificial Intelligence Safety Institute.<\/p>\n<p>An August 2024 release about those agreements indicated that the institute planned to provide feedback to both companies on \u201cpotential safety improvements to their models, in close collaboration with its partners at the UK AI Safety Institute (AISI).\u201d<\/p>\n<p>Microsoft said Tuesday in a <a href=\"https:\/\/blogs.microsoft.com\/on-the-issues\/2026\/05\/05\/advancing-ai-evaluation-with-the-center-for-ai-standards-us-and-innovation-and-the-ai-security-institute-uk\/\" target=\"_blank\" rel=\"noreferrer noopener\">blog<\/a> about the latest agreement that it, and others like it, are essential to building trust and confidence in advanced AI systems. As AI capabilities advance, it said, so too must the rigor of the testing and safeguards that underpin them.<\/p>\n<h2 class=\"wp-block-heading\" id=\"a-shift-toward-proactive-security\">A shift toward proactive security<\/h2>\n<p><a href=\"https:\/\/www.infotech.com\/profiles\/fritz-jean-louis\" target=\"_blank\" rel=\"noreferrer noopener\">Fritz Jean-Louis<\/a>, principal cybersecurity advisor at Info-Tech Research Group, said the CAISI agreements signal a shift toward proactive security for agentic AI by enabling government-led testing of advanced models before and after deployment.<\/p>\n<p>This should, he said, \u201chelp strengthen visibility into autonomous behaviors while accelerating the development of standards to mitigate risks. By combining early access, continuous evaluation, and cross-sector collaboration, the initiative pushes the industry toward security-by-design for increasingly autonomous AI systems.\u201d \u00a0<\/p>\n<p>However, added Jean-Louis, \u201cthere are a few potential hurdles to consider, for example: how would intellectual property be protected under this approach? Regardless, I believe this is a positive step for the industry.\u201d<\/p>\n<h2 class=\"wp-block-heading\" id=\"executive-order-taking-shape\">Executive order \u2018taking shape\u2019<\/h2>\n<p>Following the announcement from CAISI, a <a href=\"https:\/\/www.bloomberg.com\/news\/articles\/2026-05-06\/white-house-preps-order-to-boost-ai-security-hassett-says\" target=\"_blank\" rel=\"noreferrer noopener\">published report<\/a> on Wednesday indicated that the White House is on the verge of preparing an executive order that would see the creation of a vetting system for all new artificial intelligence models, key among them <a href=\"https:\/\/www.computerworld.com\/article\/4160021\/anthropics-latest-model-is-deliberately-less-powerful-than-mythos-and-thats-the-point.html\" target=\"_blank\">Anthropic\u2019s Mythos<\/a>.<\/p>\n<p>Bloomberg reported, \u201cthe directive is taking shape weeks after Anthropic revealed that its breakthrough Mythos model was adept at finding network vulnerabilities and could pose a global cybersecurity risk.\u201d<\/p>\n<h2 class=\"wp-block-heading\" id=\"significant-change-in-policy-direction\">Significant change in policy direction<\/h2>\n<p><a href=\"https:\/\/ca.linkedin.com\/in\/carmi\" target=\"_blank\" rel=\"noreferrer noopener\">Carmi Levy<\/a>, an independent technology analyst, said, \u201cit is patently obvious\u00a0that this week\u2019s announcement that establishes the Center for AI Standards and Innovation as the testing ground for frontier AI models is directly linked to the potential executive order that would lead to a vetting system for AI models.\u201d<\/p>\n<p>It isn\u2019t coincidental, he said, \u201cthat the announcements were made in rapid succession, and it reinforces the growing\u00a0urgency for governments in the US and elsewhere to tighten partnerships with key AI vendors to maximize AI-related security and minimize the potential for systemic risk.\u201d<\/p>\n<p>This latest flurry of activity from Washington marks a significant shift in policy direction from an administration that up until recently had been following a more laissez-faire approach to regulation, Levy pointed out.<\/p>\n<p>Concerns around Anthropic\u2019s\u00a0Claude Mythos\u00a0model, and the relative ease with which it could discover and exploit vulnerabilities in digital systems, \u201cmight have helped shift the federal government\u2019s position on AI-related regulation, particularly around the renewed push to enforce standards for AI-related deployments across government infrastructure,\u201d he said.<\/p>\n<p>AI vendors like Google, Microsoft, and xAI, Levy added, \u201cmust walk a political highwire of sorts as they balance the need to release models into the marketplace in a timely, cost-effective manner with increasingly defined rules around AI-related cybersecurity and safety. The industry can\u2019t afford a scenario where vendors themselves make up\u00a0the rules as they go along.\u201d<\/p>\n<p>At the same time, he said, the recent showdown between Anthropic and the Pentagon illustrates why the vendors might be forgiven for viewing the federal government\u2019s growing interest in AI testing and regulation with at least a certain degree of caution.<\/p>\n<p>According\u00a0 to Levy, \u201cwhile the administration\u2019s efforts to centralize testing and oversight should streamline the go-to-market process for vendors and accelerate the development of best practices around frontier model development, the political overtones of recent government-industry partnerships cannot be ignored.\u201d<\/p>\n<p><em>This article originally appeared on <a href=\"https:\/\/www.cio.com\/article\/4168122\/us-government-agency-to-safety-test-frontier-ai-models-before-release.html\" target=\"_blank\">CIO.com<\/a>.<\/em><\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>The Center for AI Standards and Innovation (CAISI), a division of the US Department of Commerce, has signed agreements with Google DeepMind, Microsoft, and xAI that would give the agency the ability to vet AI models from these organizations and others prior to their being made publicly available. According to a release from CAISI, which is part of the department\u2019s National Institute of Standards and&#8230; <\/p>\n<p class=\"more\"><a class=\"more-link\" href=\"https:\/\/newestek.com\/?p=16183\">Read More<\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"inline_featured_image":false,"footnotes":""},"categories":[1],"tags":[],"class_list":["post-16183","post","type-post","status-publish","format-standard","hentry","category-uncategorized","is-cat-link-borders-light is-cat-link-rounded"],"_links":{"self":[{"href":"https:\/\/newestek.com\/index.php?rest_route=\/wp\/v2\/posts\/16183","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/newestek.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/newestek.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/newestek.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/newestek.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=16183"}],"version-history":[{"count":0,"href":"https:\/\/newestek.com\/index.php?rest_route=\/wp\/v2\/posts\/16183\/revisions"}],"wp:attachment":[{"href":"https:\/\/newestek.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=16183"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/newestek.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=16183"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/newestek.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=16183"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}