[Congressional Bills 119th Congress] [From the U.S. Government Publishing Office] [S. 2081 Introduced in Senate (IS)] <DOC> 119th CONGRESS 1st Session S. 2081 To establish immunity from civil liability for certain artificial intelligence developers, and for other purposes. _______________________________________________________________________ IN THE SENATE OF THE UNITED STATES June 12, 2025 Ms. Lummis introduced the following bill; which was read twice and referred to the Committee on Commerce, Science, and Transportation _______________________________________________________________________ A BILL To establish immunity from civil liability for certain artificial intelligence developers, and for other purposes. Be it enacted by the Senate and House of Representatives of the United States of America in Congress assembled, SECTION 1. SHORT TITLE. This Act may be cited as the ``Responsible Innovation and Safe Expertise Act of 2025'' or the ``RISE Act of 2025''. SEC. 2. FINDINGS. Congress finds the following: (1) Artificial intelligence systems have rapidly advanced in capability and are increasingly being deployed across professional services, including healthcare, law, finance, and other sectors critical to the economy. (2) Industry leaders have publicly acknowledged the development of increasingly powerful artificial intelligence systems, with some discussing the potential for artificial general intelligence and superintelligence that could fundamentally reshape the society of the United States. (3) The current lack of clarity regarding liability for artificial intelligence errors creates uncertainty that impedes the responsible integration of these beneficial technologies into professional services and economic activity. (4) Many artificial intelligence systems operate with limited transparency regarding their capabilities, limitations, and default instructions, making it difficult for professional users to assess appropriate use cases and for legal systems to fairly allocate responsibility when errors occur. (5) Learned professionals who utilize artificial intelligence tools in serving clients have professional obligations to understand the capabilities and limitations of the tools they employ, requiring access to clear information about system specifications and performance characteristics. (6) Establishing clear standards for artificial intelligence transparency, coupled with appropriate liability frameworks, will promote responsible innovation while ensuring that the benefits and risks of artificial intelligence systems are properly understood and managed as these technologies continue to advance. (7) The development of artificial intelligence systems that may significantly impact the future of human civilization warrants a governance approach that balances innovation incentives with robust transparency requirements and appropriate allocation of responsibility among developers, professional users, and other stakeholders. SEC. 3. DEFINITIONS. In this Act: (1) Artificial intelligence.--The term ``artificial intelligence'' has the meaning given the term in section 5002 of the National Artificial Intelligence Initiative Act of 2020 (15 U.S.C. 9401). (2) Client.--The term ``client'' means a person that-- (A) engages the services of a learned professional; (B) relies upon the expertise, judgment, and advice of the learned professional; and (C) has a relationship with the learned professional that is governed by professional standards, codes of conduct, or regulations. (3) Developer.--The term ``developer'' means a person that-- (A) creates, designs, programs, trains, modifies, or substantially contributes to the creation or modification of an artificial intelligence product; (B) exercises control over the design specifications, functionality, capabilities, limitations, or intended uses of an artificial intelligence product; or (C) markets, distributes, licenses, or makes available an artificial intelligence product under their own name, brand, or trademark, regardless of whether the person creates the original underlying technology of the artificial intelligence product. (4) Error.--The term ``error'' means-- (A) any output, action, recommendation, or material omission by an artificial intelligence product that is false, misleading, fabricated, deceptive, or incomplete in a manner that a reasonable developer could foresee would cause harm; or (B) any failure of an artificial intelligence product to perform a function or task that the artificial intelligence product expressly or implicitly represents itself as capable of performing. (5) Learned professional.--The term ``learned professional'' means an individual who-- (A) possesses specialized education, training, knowledge, or skill in a profession; (B) is licensed, certified, or otherwise authorized by an appropriate Federal or State authority to practice in that profession; (C) is bound by professional standards, ethical obligations, and a duty of care to clients; and (D) exercises independent professional judgment when using tools, including artificial intelligence products, in the course of rendering professional services. (6) Model card.--The term ``model card'' means a publicly available technical document in which a developer describes, consistent with industry standards and as rigorously as or more rigorously than industry peers, the training data sources, evaluation methodology, performance metrics, intended uses, limitations, and risk mitigations, including detection, evaluation, management, and safeguards against errors, of an artificial intelligence product. (7) Model specification.--The term ``model specification''-- (A) means the text or other configuration instructions of an artificial intelligence product-- (i) supplied by a developer; (ii) that establish the intended base behavior, tone, constraints, or goals of the artificial intelligence product; and (iii) that materially influence the outputs of the artificial intelligence product across users or sessions, including the system prompt provided to the model before engaging with user queries; and (B) includes-- (i) the system prompt and any other text or images that the artificial intelligence product receives that are not visible to the end user; (ii) any constitution or analogous guiding document used when training or fine-tuning of an artificial intelligence product, including in automated schemes in which an artificial intelligence system trains another artificial intelligence system; and (iii) the instructions, rubrics, or other guidance provided to human raters or evaluators of an artificial intelligence product the feedback of whom is used to train or fine-tune the artificial intelligence product. SEC. 4. CONDITIONAL IMMUNITY FROM CIVIL LIABILITY FOR ARTIFICIAL INTELLIGENCE DEVELOPERS. (a) Safe Harbor Eligibility.--A developer shall be immune from civil liability for errors generated by an artificial intelligence product when used by a learned professional in the course of providing professional services to a client if the developer-- (1) prior to deployment of the artificial intelligence product, publicly releases and continuously maintains-- (A) the model card for the artificial intelligence product; and (B) the model specification for the artificial intelligence product, which may include redactions-- (i) only relating to information that would reveal trade secrets unrelated to the safety of the artificial intelligence product; and (ii) only if the developer furnishes contemporaneously with each redaction a written justification for the redaction identifying the basis for withholding the information as a trade secret; and (2) provides clear and conspicuous documentation to learned professionals describing the known limitations, failure modes, and appropriate domains of use for the artificial intelligence product. (b) Scope of Immunity.--The immunity provided under subsection (a) shall be conferred to a developer only for acts or omissions that do not constitute recklessness or willful misconduct by the developer. (c) Duty To Update.--Immunity under subsection (a) relating to an artificial intelligence product shall not apply to a developer-- (1) that does not update the model card, model specification, and documentation with respect to the artificial intelligence product as described in subsection (a)(1) by the date that is 30 days after the date on which the developer-- (A) deploys a new version of the artificial intelligence product; or (B) discovers a new and material failure mode affecting the artificial intelligence product; and (2) of which the failure to make an update described in paragraph (1) by the applicable date described in that paragraph proximately causes a harm occurring after that date. (d) Preemption.-- (1) Express preemption.--This section shall apply to any claim arising under State law against a developer for an error arising from the use of an artificial intelligence product by a learned professional in providing professional services if the developer is immune from civil liability under subsection (a). (2) Claims not preempted.--Nothing in this section shall apply to a claim arising under State law against a developer based on fraud, knowing misrepresentation, or conduct outside the scope of professional use of an artificial intelligence product by a learned professional. SEC. 5. PRESERVATION OF OTHER IMMUNITIES AND PRIVILEGES. Nothing in this Act shall be construed to affect any immunity from civil liability established by Federal or State law or available at common law that is not related to the immunity established under section 4(a). SEC. 6. EFFECTIVE DATE; APPLICABILITY. This Act-- (1) shall take effect on December 1, 2025; and (2) shall apply to acts or omissions occurring on or after the date described in paragraph (1). <all>