Codify — Article

Reliable Artificial Intelligence Research Act of 2025: DHS to run AI prize competitions

Mandates DHS-run prize challenges to advance interpretability and adversarial robustness for commercially used AI, authorizes $10M and cross-agency consultation to seed standards for high‑risk AI.

The Brief

The bill requires the Secretary of Homeland Security to launch prize competitions—under the Stevenson‑Wydler prize authority—to accelerate research on two specific areas: interpretability for commercially available or widely used AI products, and adversarial robustness for models intended for at least one high‑impact, high‑risk application. Each competition must begin within 270 days of enactment, may be phased, and must be designed with defined evaluation criteria and cross‑agency consultation.

Why it matters: prize competitions are a low‑touch federal tool to catalyze research, build benchmarks, and potentially seed interoperable standards that could influence procurement and industry practice. The legislation also requires DHS to report evaluation findings to congressional homeland security committees and authorizes $10 million for implementation across FY2026–2030—small but potentially catalytic funding tied to standard‑setting outcomes rather than grants or contracts alone.

At a Glance

What It Does

The bill directs DHS to run at least one prize competition for interpretability and at least one for adversarial robustness using the Stevenson‑Wydler prize authority, each to begin within 270 days of enactment. Competitions may be multi‑phased (frameworks, model submissions, red‑teaming), use tailored evaluation criteria, and be administered via contracts or cooperative agreements.

Who It Affects

Affected parties include AI developers and vendors of commercially used models, federal agencies that deploy or regulate high‑risk AI, research institutions and toolmakers focused on interpretability and robustness, and private firms that provide red‑teaming or evaluation services. DHS’s contractors and the agencies consulted (Commerce, NIST, NSF, National Cyber Director) will also be operationally engaged.

Why It Matters

The competitions are designed to produce reusable artifacts—interpretability frameworks, robust model baselines, red‑team playbooks—that can inform government procurement standards and industry best practices. Because the bill ties evaluation to applicability in high‑risk use cases and the potential to create standards, winning entries could shape what counts as ‘acceptable’ interpretability and robustness in sensitive applications.

More articles like this one.

A weekly email with all the latest developments on this topic.

Unsubscribe anytime.

What This Bill Actually Does

The Reliable Artificial Intelligence Research Act directs the Secretary of Homeland Security to use an established federal prize mechanism to accelerate two narrow but consequential strands of AI research: making models more interpretable to humans and making them resistant to adversarial attacks. For interpretability, DHS must design competitions that produce broadly applicable frameworks or demonstrable interpretable models relevant to widely used commercial systems.

For robustness, competitions must target capable models that show resistance to adversarial inputs in at least one identified high‑impact, high‑risk application.

The bill sets concrete procedural guardrails. DHS must start each competition within 270 days of enactment, consult specified agencies and industry experts in designing the contests, and can split competitions into phases—examples in the text include framework submissions, model submissions, and red‑teaming.

Evaluation criteria must weigh generalizability, practical value for high‑risk uses, and the likelihood that submissions will inform standards in government or industry. DHS may contract with outside entities—nonprofits, for‑profits, or state/tribal bodies—to run the contests.After the first competition concludes, DHS must report to the Senate Committee on Homeland Security and Governmental Affairs and the House Homeland Security Committee within 180 days with an evaluation of what the competitions produced, gaps identified, and suggested congressional actions.

Congress has authorized $10 million to DHS for FY2026–2030 to carry out the statute, leaving DHS to prioritize prize design, outreach, and administration within that funding envelope.Operationally, the statute focuses on deliverables that can be reused: interpretability frameworks that are broadly applicable; robust model prototypes and documented red‑teaming methods; and evaluations that explicitly tie submissions to high‑risk, high‑value use cases. The combination of short start timelines, cross‑agency consultation, and a narrowly scoped budget pushes DHS toward targeted, leverage‑oriented competitions rather than large sustained research programs.

The Five Things You Need to Know

1

The Secretary of Homeland Security must commence at least one interpretability prize competition and at least one adversarial‑robustness competition within 270 days of enactment, using the Stevenson‑Wydler prize authority.

2

Competitions may be multi‑phased (frameworks, model submissions, red‑teaming) and may open different phases to distinct contestant pools to capture both basic research and applied model development.

3

Evaluation criteria must prioritize broadly applicable interpretability or robustness principles, practical value in high‑risk, high‑value use cases, and the potential for submissions to create government or industry standards.

4

DHS must consult Commerce, NIST, the National Cyber Director, NSF, industry experts, and—for robustness competitions—heads of agencies with high‑risk AI expertise when designing contests.

5

Congress authorized $10,000,000 for FY2026–2030 and requires DHS to report to the Senate and House homeland security committees within 180 days after the first competition concludes with findings and suggested next steps.

Section-by-Section Breakdown

Every bill we cover gets an analysis of its key sections. Expand all ↓

Section 1

Short title

Names the statute the "Reliable Artificial Intelligence Research Act of 2025." This is the legal citation used throughout implementation documents and contracts; it signals congressional intent to frame the work as research/standards‑oriented rather than regulatory.

Section 2

Definitions

Provides operative definitions for adversarial robustness, artificial intelligence (by reference to the National AI Initiative Act), interpretability, red‑teaming, and Secretary (DHS). These definitions constrain the scope of competitions—for example, interpretability must connect to human understanding of model decision processes; robustness focuses on resisting attacks while preserving privacy and integrity. Referencing the NAIIA definition of AI imports that statute's scope and avoids re‑litigating a separate AI definition here.

Section 3

Interpretability prize competition

Requires DHS to launch at least one prize competition focused on interpretability relevant to commercially available or widely used AI products. DHS must consult named agencies and industry experts, set structure and evaluation criteria aligned to advancing broadly applicable interpretability principles, and may phase contests into framework submissions, interpretable models, and novel basic research. DHS may use contracts or cooperative agreements to administer the competition, allowing the agency to leverage private‑sector organizers or academic partners.

3 more sections
Section 4

Adversarial robustness prize competition

Mirrors Section 3 but targets adversarial robustness, requiring competitions to produce capable models designed to be robust in at least one high‑impact, high‑risk application. The consultation list is broader (including agency heads with domain expertise), and phases explicitly include red‑teaming. Evaluation must measure advancement of broadly applicable robustness principles and reductions in adversarial risk for high‑value use cases.

Section 5

Tracking and reporting

Requires DHS to report to the Senate Homeland Security and Governmental Affairs Committee and the House Homeland Security Committee within 180 days after the first competition ends. The report must evaluate how results inform interpretability and robustness, identify research gaps, and recommend congressional actions. The reporting requirement creates an explicit feedback loop that can trigger further legislative or appropriations responses based on contest outcomes.

Section 6

Appropriations

Authorizes $10 million for DHS to carry out the Act across FY2026–2030. This is an authorization, not a direct appropriation; actual execution will depend on appropriations. The modest size relative to major ML research costs will force DHS to use prize structuring, partnerships, and contracting strategically to maximize leverage.

At scale

This bill is one of many.

Codify tracks hundreds of bills on Technology across all five countries.

Explore Technology in Codify Search →

Who Benefits and Who Bears the Cost

Every bill creates winners and losers. Here's who stands to gain and who bears the cost.

Who Benefits

  • Academic researchers and interpretability toolmakers — competitions create directed funding opportunities, public benchmarks, and visibility for methods that can scale to commercially used systems.
  • Federal procurement and program offices that deploy high‑risk AI — the work can produce testable standards or baselines to inform procurement requirements and risk assessments.
  • Independent red‑team vendors and security researchers — red‑teaming phases create paid, structured demand for adversarial testing services and playbooks that can be commercialized or adopted by agencies.
  • Startups and small firms building explainability or robustness tooling — successful entries can translate into commercial credibility and potential government customer recognition.
  • Standards bodies and NIST — DHS consultations and competition results may supply concrete artifacts and candidate metrics useful for standard development.

Who Bears the Cost

  • Department of Homeland Security — will bear program administration, design, and oversight responsibilities within a constrained $10M authorization and will need to coordinate widely across agencies.
  • Participating firms and research teams — preparing high‑quality submissions (models, red‑team reports, reproducible code) can be resource‑intensive and may expose intellectual property or proprietary model details.
  • Taxpayers and appropriators — Congress must convert the authorization into appropriations; sustained standard‑setting or follow‑on work will likely require additional funding.
  • Smaller labs and underfunded researchers — prize formats often advantage well‑resourced teams able to iterate quickly on models and to field red‑teams, potentially skewing results toward incumbents.
  • Consulted federal agencies — Commerce, NIST, NSF, and others will need to allocate staff time and expertise to design and evaluate competitions without explicit new funding in the text.

Key Issues

The Core Tension

The central dilemma is whether a prize competition run by a security‑focused agency can simultaneously produce open, reusable interpretability artifacts and robust defenses without exposing sensitive model details or privileging incumbents: the statute aims to accelerate public‑good standards while relying on competitive incentives that favor well‑resourced actors and may require disclosure of vulnerabilities and proprietary techniques.

Two implementation challenges are prominent. First, the statute combines an openness goal (interpretable methods and public frameworks) with a security goal (robustness and red‑teaming) that can point in opposite directions.

Publicly releasing model internals or detailed red‑team findings accelerates research and standards work but can also reveal exploitable flaws or proprietary techniques. The bill leaves handling of sensitive disclosures to program design and contracts but gives no specific guidance on IP, data sensitivity, or secure evaluation environments.

Second, the funding and institutional placement create tradeoffs. DHS is not a primary federal science funder for AI; it will rely on cross‑agency consultation and external contractors to attract top research talent.

The authorized $10 million over five years is modest relative to leading‑edge machine learning costs, which raises questions about whether contests will produce cutting‑edge, production‑grade models or instead yield proof‑of‑concepts and frameworks. Finally, the bill ties success metrics to ‘‘likelihood’’ of shaping standards and applicability to high‑risk use cases—both inherently judgmental tests that place considerable discretion in DHS and its evaluation panels, increasing the importance of transparent scoring rubrics and conflict‑of‑interest safeguards.

Try it yourself.

Ask a question in plain English, or pick a topic below. Results in seconds.