AIモデルの事前審査に向けた自主的なアプローチ:タイヤを蹴る

#Tech

AIモデルの事前審査に向けた自主的なアプローチ:タイヤを蹴る

トランプ政権は、最先端AIモデルに対する「審査システム」の導入を検討しています。

これは、AI開発者がモデルを政府に先行公開し、サイバー攻撃能力を持つモデルのリスクを評価し、対策を講じることを目的としたものです。

強制的な審査システムを構築する法的根拠は不明確ですが、自主的な「タイヤを蹴る」テスト期間を設け、モデルの評価結果をCAISIを通じて共有し、CISAがセキュリティ対策を支援することが可能です。

Anthropic社のMythosモデルの事例や、AIツールが社会に浸透する前にセキュリティ対策が十分に行われないことへの懸念から、事前評価の重要性が浮き彫りになっています。

強制的な審査の代わりに、自主的な評価を通じてリスクを事前に把握し、対応策を講じることで、広範な社会への影響を軽減できると考えられています。

トランプ政権が、最先端AIモデル(フロンティアAI)に対する「審査システム」の創設を検討していることが明らかになりました。これは、AI開発企業が自発的に政府にモデルを共有し、潜在的なリスクを事前に検証するという、一種の「試運転」を促す仕組みです。AIの急速な進化に伴い、サイバーセキュリティ上の脅威が増大する中、この自主的な検証プロセスがどのような形で導入されるのかが注目されています。

AIがもたらす新たなサイバー脅威

Anthropic社は、最新モデル「Mythos」がソフトウェアの脆弱性を特定・悪用できる能力を持つため、当初のリリースを見送りました。同社は、ホワイトハウスや連邦政府と交渉し、選定された民間関係者に対してモデルを先行提供しました。この判断は単なるPR戦略と見なされることもありましたが、第三者によるテストの結果、Mythosが人間が20時間かかるような32ステップの企業ネットワーク攻撃を信頼性をもって監督できることが判明しています。

OpenAI社も同様のサイバー能力を持つモデルを開発しており、専門家たちは、中小企業や非営利団体など、セキュリティ対策が十分でない社会全体が、これらのAIツールによって突然脆弱になることを懸念しています。AIの進化速度に対し、防御側の準備が追いついていないという深刻な非対称性が生じている状況です。

自主的検証のメリットと課題

この自主的な「試運転」制度は、AI開発企業が、自社のモデルが重要システムや公共の福祉を脅かす可能性があることを知りながらも市場に投入した際に、世論からの反発を避ける助けになると考えられています。企業は、政府や専門機関(CAISIなど)にモデルを共有することで、リスクを事前に洗い出し、対策を講じることができます。

しかし、この仕組みが機能するためには、サイバーセキュリティ機関(CISA)などが、地方自治体や民間企業を含む幅広い関係者に対して、必要なセキュリティ対策を支援するための資金提供を行う必要があります。自主的な取り組みであるため、全ての開発者が参加するとは限らないという課題も残ります。

強制的な審査の法的ハードル

トランプ政権が、AIモデルの審査を強制的に義務付ける法的権限を持つかについては、非常に判断が難しい状況です。いくつかの国家安全保障関連法が候補として挙げられていますが、それらをAIモデルの審査に適用するには、法律の解釈を大きく押し広げる必要があり、裁判所が認める可能性は低いと見られています。

特に「国防生産法(DPA)」は、行政権の拡大に利用されてきた経緯がありますが、その本来の目的は「緊急時に国家防衛に必要な資源を民間産業に強制的に生産させること」にあります。AIモデルの審査という目的に当てはめるには、法律の枠組みが合致しない可能性が高いと分析されています。

結論

AIの脅威が増す中、政府による規制導入の動きは注目されています。現時点では、企業が自発的にリスクを共有する「自主的な検証」が現実的な道筋と見られますが、この仕組みを実効性のあるものにするためには、法的な枠組みの明確化と、広範な社会へのセキュリティ支援が不可欠です。

原文の冒頭を表示(英語・3段落のみ)

The Trump administration is weighing the creation of a “review system” for frontier AI models. According to the New York Times, in this proposed approach, AI labs would provide the federal government with “first access” to “get ahead” of models with significant cyber capabilities, presumably such as Anthropic’s Mythos. It’s unclear what legal authority would allow the president to accomplish these goals—specifically, mandating labs to undergo a vetting process and then sharing any essential information related to countering any detected risks with other parts of the government.However, existing authorities would allow for a voluntary “kick the tires” testing period. Labs could opt to share models with materially new capabilities with the Center for AI Standards and Innovation (CAISI), which is housed within the Commerce Department’s National Institute of Standards and Technology; the director of the Cybersecurity and Infrastructure Security Agency (CISA) could then fund an effort to help a broad set of actors—including local, state, and federal actors as well as other public and private entities—take any necessary cybersecurity precautions. This would help labs avoid popular backlash for knowingly introducing models that may threaten critical systems and public well-being and perhaps subvert more onerous, formal requirements.Cyber Threats Posed by AIAnthropic opted not to release their latest model, Mythos, because of its ability to identify and exploit software vulnerabilities. Instead, Anthropic made the model available to a select group of private stakeholders and, following negotiations with the White House, the federal government, to take any necessary precautions. Though some dismissed that decision as a PR move, testing from third parties validated Anthropic’s findings; by way of example, the U.K. AI Security Institute determined that Mythos could oversee a 32-step corporate network attack with some degree of reliability—a process that would take humans 20 hours. OpenAI subsequently developed a model with similar cyber capabilities.Cybersecurity experts fear that large swaths of society—such as private- and public-sector entities with mediocre cybersecurity plans—may be caught flat-footed when other such tools become available. Small businesses and nonprofits, for example, “lack the skills and resources to address these challenges before their systems are compromised.” State and local governments may similarly be ill-equipped to take timely measures as AI tools continue to become more sophisticated. This is especially concerning given that local election officials report a dearth of state and federal support for countering emerging cyber threats.The result is a widening asymmetry: Frontier labs increasingly understand what their models can do weeks or months before the institutions tasked with defending against misuse have any meaningful opportunity to prepare.Can the President Mandate a Vetting Process?Whether the president has the legal authority to mandate a vetting process is hard to assess without the Trump administration specifying its own understanding of the law. A preliminary review of national security provisions returns a potential shortlist: the Defense Production Act (DPA), the International Emergency Economic Powers Act (IEEPA), and the Communications Act of 1934. For sake of brevity, the latter two can be dismissed fairly easily. Reliance on those acts to subject U.S. companies to a government “review system” would involve stretched interpretations of the laws, which courts would likely not condone.The DPA is also an unlikely stable legal basis for the administration’s plan. President Biden leaned on the DPA to require AI model testing under an executive order he issued in 2023; President Trump rescinded that order in 2025. The breadth of the DPA’s terms have made it a recurring vehicle for expansive actions by the executive branch. Pursuant to the DPA, a president may use an “array of authorities to shape national defense preparedness programs and to take appropriate steps to maintain and enhance the domestic industrial base.” However, numerous investigations of the DPA have pointed out that those authorities have limits.When the DPA was enacted in 1950, its original purpose “was to ensure that the federal government could compel private industry to produce strategically necessary resources to meet the needs of national defense during an emergency,” according to Ashley Mehra. While the law has subsequently been amended and stretched by creative uses of its vague authorities, it is unlikely to fit the administration’s unique legal needs here. The president will be hard pressed to find a clear hook for compelled vetting of an AI model with minimal direct connections to a specific, ongoing national defense effort or domestic industrial base consideration in the DPA. Title I of the DPA empowers the president to direct private parties to prioritize and accept contracts necessary for national defense. Title III enables the president to incentivize the production of certain critical materials and goods, such as with loan guarantees and grants. Title VII includes a range of authorities, including the ability to establish voluntary arrangements among private actors that might otherwise run afoul of antitrust laws, as well as to gather information from private entities. More specifically, the Department of Commerce may rely on Title VII to “conduct assessments of domestic industrial base capabilities.”Use of Title VII to govern frontier AI models runs counter to the marginal role intended for this part of the DPA. In contrast to Titles I and III, Title VII amounts to a “potpourri” of provisions meant to assist with administration of the act more so than to afford expansive powers. A more common understanding of the information-gathering power afforded therein is the authority to “obtain information from industry and firms, including through testimony or by inspecting their books, records and properties.” Such an inquiry would serve the purpose of identifying any weak points in supply chains that may be relevant to the nation’s readiness for war and related emergencies.Compelled disclosure of a model to the federal government does not fit neatly into any DPA authorities.The Legal Path to a Voluntary “Kick the Tires” PeriodThough it seems unlikely that the president can force AI labs to participate in a review system, he does not need to if the underlying goal is to translate the results of CAISI’s voluntary evaluations of model capabilities into general cyber readiness assistance. The leading labs already make their models available to CAISI for rigorous testing on a voluntary basis. Given that they have a vested interest in avoiding the following headline: “AI Lab Bypasses Federal Testing; Cyberattacks Proliferate,” they may agree for such testing to occur two or three weeks prior to generally deploying their model.A more evocative hypothetical stresses this point. Imagine that a leading lab releases a model with notable cyber capabilities days before the November elections. Political actors may allege that local and state officials likely had their election systems undermined by bad actors using the latest AI model. It would be hard to dismiss such a claim under the status quo. One could easily foresee reports on “Model ____ Blamed for Cyberattacks; Election Results Contested.” Such headlines would become far less likely if a short-term “kick the tires” period became standard practice.Suppose instead that prior to that model’s release, CAISI concludes—based on publicly disclosed, objective evaluations that return clearly established threat models—that the model would indeed expose critical infrastructure, local, state, tribal, and federal actors, and private entities to cyber threats, they can immediately share that information with CISA. The CISA director would then need to evaluate whether a “specific significant incident is likely to occur imminently” in order to trigger additional authorities under the Homeland Security Act.Significant incident refers to “an incident or a group of related incidents that results, or is likely to result, in demonstrable harm to—(i) the national security interests, foreign relations, or economy of the United States; or (ii) the public confidence, civil liberties, or public health and safety of the people of the United States.” Mass deployment of a model with heightened cyber capabilities seems likely to qualify. As noted above, Mythos is capable of hacks that require 20 hours of human work. If a model of a similar capability was in the hands of even a couple dozen bad actors, there would likely be widespread economic harm and, perhaps, threats to the public health and safety.The CISA director could then use the Cyber Response and Recovery Fund to help a range of stakeholders—public and private—with “vulnerability assessments and mitigation; technical incident mitigation; malware analysis; analytic support; threat detection and hunting; and, network protections.”Of course, the biggest “if” here is whether labs would agree to this “kick the tires” period. Labs have reasons to say yes.A specific, time-bound testing window run by CAISI is preferable to the alternatives on the table—whether that means a mandatory review regime, which would likely run afoul of the law, or the reputational fallout of a post-deployment incident traced back to capabilities the lab knew about but did not flag.CISA, for its part, would need to commit to handling shared model information with the same care it extends to vulnerability disclosures from private security researchers, lest labs conclude that cooperation invites leakage rather than partnership.Such an approach would not require new legislation, nor would it require the administration to stretch the DPA past its breaking point. It requires only that the relevant agencies use the authorities they already have, and that the labs recognize that a voluntary framework is the one most likely to keep a mandatory one off the table.

※ 著作権に配慮し、引用は冒頭3段落までです。続きは元記事をご覧ください。

元記事を読む ↗