[Congressional Bills 119th Congress]
[From the U.S. Government Publishing Office]
[S. 2938 Introduced in Senate (IS)]
<DOC>
119th CONGRESS
1st Session
S. 2938
To require the Secretary of Energy to establish the Advanced Artificial
Intelligence Evaluation Program, and for other purposes.
_______________________________________________________________________
IN THE SENATE OF THE UNITED STATES
September 29, 2025
Mr. Hawley (for himself and Mr. Blumenthal) introduced the following
bill; which was read twice and referred to the Committee on Commerce,
Science, and Transportation
_______________________________________________________________________
A BILL
To require the Secretary of Energy to establish the Advanced Artificial
Intelligence Evaluation Program, and for other purposes.
Be it enacted by the Senate and House of Representatives of the
United States of America in Congress assembled,
SECTION 1. SHORT TITLE.
This Act may be cited as the ``Artificial Intelligence Risk
Evaluation Act of 2025''.
SEC. 2. SENSE OF CONGRESS; PURPOSES.
(a) Sense of Congress.--It is the sense of Congress that rapidly
advancing artificial intelligence capabilities present both
opportunities and significant risks to national security, public
safety, economic competitiveness, civil liberties, and healthy labor
and other markets, and that, as artificial intelligence advances toward
human-level capabilities in virtually all domains, the United States
must establish a secure testing and evaluation program to generate
data-driven options for managing emerging risks.
(b) Purposes.--The purposes of the program established under this
Act are to provide Congress with the empirical data, lessons, and
insights necessary for Federal oversight of artificial intelligence to
ensure that regulatory decisions are made on the basis of empirical
testing, and to enable Congress to safeguard American citizens.
SEC. 3. DEFINITIONS.
In this Act:
(1) Advanced artificial intelligence system.--
(A) In general.--Subject to subparagraph (B), the
term ``advanced artificial intelligence system'' means
an artificial intelligence system that was trained
using a quantity of computing power greater than 10\26\
integer or floating-point operations.
(B) Alternate meaning.--The Secretary may, by a
rule, propose a new definition of the term ``advanced
artificial intelligence system'' to replace the
definition in subparagraph (A), which new definition
shall not go into effect until the Secretary submits
the rule to Congress and a joint resolution approving
the rule is enacted into law.
(2) Adverse AI incident.--The term ``adverse AI incident''
means an incident relating to an artificial intelligence system
that involves--
(A) a loss-of-control scenario;
(B) a risk of weaponization by a foreign adversary,
a foreign terrorist organization, or another adversary
of the United States Government;
(C) a threat to the safety or reliability of
critical infrastructure (as defined in subsection (e)
of the Critical Infrastructures Protection Act of 2001
(42 U.S.C. 5195c(e)));
(D) a significant erosion of civil liberties,
economic competition, and healthy labor markets;
(E) scheming behavior; or
(F) an attempt to carry out an incident described
in subparagraphs (A) through (E).
(3) Artificial intelligence; AI.--The term ``artificial
intelligence'' or ``AI'' means technology that enables a device
or software--
(A) to make--for a given set of human-defined
objectives--predictions, recommendations, or decisions
influencing real or virtual environments; and
(B) to use machine and human-based inputs--
(i) to perceive real and virtual
environments;
(ii) to abstract such perceptions into
models through analysis in an automated manner;
and
(iii) to use model inference to formulate
options for information or action.
(4) Artificial intelligence system; AI system.--The term
``artificial intelligence system'' or ``AI system'' means a
particular model, program, or tool within the field of
artificial intelligence.
(5) Artificial superintelligence.--
(A) In general.--The term ``artificial
superintelligence'' means artificial intelligence that
exhibits, or can easily be modified to exhibit, all of
the characteristics described in subparagraph (B).
(B) Characteristics described.--The characteristics
referred to in subparagraph (A) are the following:
(i) The AI can enable a device or software
to operate autonomously and effectively for
long stretches of time in open-ended
environments and in pursuit of broad
objectives.
(ii) The AI can enable a device or software
to match or exceed human cognitive performance
and capabilities across most domains or tasks,
including those related to decisionmaking,
learning, and adaptive behaviors.
(iii) The AI can enable a device or
software to potentially exhibit the capacity to
independently modify or enhance its own
functions in ways that could plausibly
circumvent human control or oversight, posing
substantial and unprecedented risks to
humanity.
(6) Computing power.--The term ``computing power'' means
the processing power and other electronic resources used to
train, validate, deploy, and run AI algorithms and models.
(7) Covered advanced artificial intelligence system
developer.--The term ``covered advanced artificial intelligence
system developer'' means a person that designs, codes,
produces, owns, or substantially modifies an advanced
artificial intelligence system for use in interstate or foreign
commerce, including by taking steps to initiate a training run
of the advanced artificial intelligence system.
(8) Deploy.--The term ``deploy'' means an action taken by a
covered advanced artificial intelligence system developer to
release, sell, or otherwise provide access to an advanced
artificial intelligence system outside the custody of the
developer, including by releasing an open-source advanced
artificial intelligence system.
(9) Foreign adversary.--The term ``foreign adversary''
means a foreign adversary (as defined in section 791.2 of title
15, Code of Federal Regulations) (or successor regulations)
that is included on the list in section 791.4(a) of that title
(or successor regulations).
(10) Foreign terrorist organization.--The term ``foreign
terrorist organization'' means a foreign entity designated as a
foreign terrorist organization by the Secretary of State under
section 219 of the Immigration and Nationality Act (8 U.S.C.
1189).
(11) Interstate or foreign commerce.--The term ``interstate
or foreign commerce'' has the meaning given the term in section
921(a) of title 18, United States Code.
(12) Loss-of-control scenario.--The term ``loss-of-control
scenario'' means a scenario in which an artificial intelligence
system--
(A) behaves contrary to its instruction or
programming by human designers or operators;
(B) deviates from rules established by human
designers or operators;
(C) alters operational rules or safety constraints
without authorization;
(D) operates beyond the scope intended by human
designers or operators;
(E) pursues goals that are different from those
intended by human designers or operators;
(F) subverts oversight or shutdown mechanisms; or
(G) otherwise behaves in an unpredictable manner so
as to be harmful to humanity.
(13) Program.--The term ``program'' means the Advanced
Artificial Intelligence Evaluation Program established under
section 5.
(14) Scheming behavior.--The term ``scheming behavior''
means behavior by an AI system to deceive human designers or
operators, including by--
(A) hiding its true capabilities and objectives; or
(B) attempting to subvert oversight mechanisms or
shutdown mechanisms.
(15) Secretary.--The term ``Secretary'' means the Secretary
of Energy.
SEC. 4. OBLIGATION TO PARTICIPATE; ENFORCEMENT AND PENALTIES.
(a) In General.--Each covered advanced artificial intelligence
system developer shall--
(1) participate in the program; and
(2) provide to the Secretary, on request, materials and
information necessary to carry out the program, which may
include, with respect to the advanced artificial intelligence
system of the covered advanced artificial intelligence system
developer--
(A) the underlying code of the advanced artificial
intelligence system;
(B) data used to train the advanced artificial
intelligence system;
(C) model weights or other adjustable parameters
for the advanced artificial intelligence system;
(D) the interface engine or other implementation of
the advanced artificial intelligence system; and
(E) detailed information regarding the training,
model architecture, or other aspects of the advanced
artificial intelligence system.
(b) Prohibition on Deployment.--No person may deploy an advanced
artificial intelligence system for use in interstate or foreign
commerce unless that person is in compliance with subsection (a).
(c) Penalty.--A person that violates subsection (a) or (b) shall be
fined not less than $1,000,000 per day of the violation.
SEC. 5. ADVANCED ARTIFICIAL INTELLIGENCE EVALUATION PROGRAM.
(a) In General.--Not later than 90 days after the date of enactment
of this Act, the Secretary shall establish an Advanced Artificial
Intelligence Evaluation Program within the Department of Energy.
(b) Activities.--The program shall--
(1) offer standardized and classified testing and
evaluation of advanced AI systems to systematically collect
data on the likelihood of adverse AI incidents for a given
advanced AI system;
(2) implement testing protocols that match or exceed
anticipated real-world AI jailbreaking techniques, including
adversarial testing by red teams with expertise comparable to
sophisticated malicious actors;
(3) to the extent feasible, establish and facilitate
classified, independent third-party assessments and blind model
evaluations to maintain transparency and reliability;
(4) provide participating entities with a formal report
based on testing outcomes that clearly identifies evaluated
risks and safety measures;
(5) develop recommended containment protocols, contingency
planning, and mitigation strategies informed by testing data to
address identified risks;
(6) inform the creation of evidence-based standards,
regulatory options, guidelines, and governance mechanisms based
on data collected from testing and evaluations;
(7) assist Congress in determining the potential for
controlled AI systems to reach artificial superintelligence,
exceed human oversight or operational control, or pose
existential threats to humanity by providing comprehensive
empirical evaluations and risk assessments; and
(8) develop proposed options for regulatory or governmental
oversight, including potential nationalization or other
strategic measures, for preventing or managing the development
of artificial superintelligence if artificial superintelligence
seems likely to arise.
(c) Plan for Permanent Framework.--
(1) In general.--Not later than 360 days after the date of
enactment of this Act, the Secretary shall submit to Congress a
detailed recommendation for Federal oversight of advanced
artificial intelligence systems, drawing directly upon
insights, empirical data, and lessons learned from the program.
(2) Contents.--The plan submitted under paragraph (1)
shall--
(A) summarize and analyze outcomes from testing,
identifying key trends, capabilities, potential risks,
and system behaviors such as weaponization potential,
self-replication capabilities, scheming behaviors,
autonomous decisionmaking, and automated AI development
capabilities;
(B) recommend evidence-based standards,
certification procedures, licensing requirements, and
regulatory oversight structures specifically informed
by testing and evaluation data, ensuring alignment
between identified risks and regulatory responses;
(C) outline proposals for automated and continuous
monitoring of AI hardware usage, computational resource
inputs, and cloud-computing deployments based on
observed relationships between those factors and AI
system performance or emergent capabilities;
(D) propose adaptive governance strategies that
account for ongoing improvements in algorithmic
efficiency and system capabilities, ensuring that
regulatory frameworks remain relevant and effective as
AI technology advances;
(E) suggest revisions with respect to Federal
oversight or resourcing, such as a new office within an
existing agency, a new agency, or additional funding,
that may be necessary to develop and administer a
permanent framework for oversight of advanced
artificial intelligence systems; and
(F) provide comprehensive evaluations regarding the
potential for tested AI systems to exceed human
oversight, approach artificial superintelligence,
threaten economic competition (including in labor
markets), undermine civil liberties, and pose
existential risks to humanity, including clearly
articulated options for regulatory or governmental
oversight measures to address scenarios of imminent
concern identified through testing.
(3) Updates.--Not less frequently than once every year for
the duration of the program, the Secretary shall--
(A) update the plan submitted under paragraph (1)
with new insights, data, and lessons from the program;
and
(B) submit the updated plan to Congress.
(d) Sunset.--The program shall terminate on the date that is 7
years after the date of enactment of this Act, unless renewed by
Congress.
<all>