EXCLUSIVE
Artificial intelligence

New GOP bill would protect AI companies from lawsuits if they offer transparency

This version of Cynthia Lummis Ai Civil Liability Protections Developers Transparency Rcna212131 - Technology and Innovation | NBC News Clone was adapted by NBC News Clone to help readers digest key facts more efficiently.

First to NBC News: Sen. Cynthia Lummis' legislation seeks to address the legal gray area of exactly who is liable when AI is used in sensitive medical, legal or financial situations.
Sen. Cynthia Lummis
Sen. Cynthia Lummis, R-Wyo., arrives for a vote at the Capitol on May 14.Bill Clark / CQ-Roll Call via Getty Images file

Sen. Cynthia Lummis, R-Wyo., is introducing legislation Thursday that would shield artificial intelligence developers from an array of civil liability lawsuits provided they meet certain disclosure requirements.

Lummis’ bill, the Responsible Innovation and Safe Expertise Act, seeks to clarify that doctors, lawyers, financial advisers, engineers and other professionals who use AI programs in their decision-making retain legal liability for any errors they make — so long as AI developers publicly disclose how their systems work.

“This legislation doesn’t create blanket immunity for AI — in fact, it requires AI developers to publicly disclose model specifications so professionals can make informed decisions about the AI tools they choose to utilize,” Lummis, a member of the Commerce Committee, said in a statement first shared with NBC News. “It also means that licensed professionals are ultimately responsible for the advice and decisions they make. This is smart policy for the digital age that protects innovation, demands transparency, and puts professionals and their clients first.”

Lummis’ office touted the bill as the first piece of federal legislation that offers clear guidelines for AI liability in a professional context. The measure would not govern liability for other AI elements, such as self-driving vehicles, and it would not provide immunity when AI developers act recklessly or willfully engage in misconduct.

“AI is transforming industries — medicine, law, engineering, finance — and becoming embedded in professional tools that shape critical decisions,” her office said in a release. “But outdated liability rules discourage innovation, exposing developers to unbounded legal risk even when trained professionals are using these tools.”

Exactly who is liable when AI is used in sensitive medical, legal or financial situations is a bit of a gray area, with some states seeking to enact their own standards.

The House-passed “One Big Beautiful Bill,” which is advancing through Congress and supported by President Donald Trump, includes a provision that would ban states from enacting any AI regulations for 10 years. Senate Republicans last week proposed changing the provision to instead block federal funding for broadband projects to states that regulate AI.

Both Democratic and Republican state officials have criticized the effort to prohibit state-level regulations over the next decade, while AI executives have argued that varying state laws would stifle industry growth when the United States is in stiff competition with countries like China.

×
AdBlock Detected!
Please disable it to support our content.

Related Articles

Donald Trump Presidency Updates - Politics and Government | NBC News Clone | Inflation Rates 2025 Analysis - Business and Economy | NBC News Clone | Latest Vaccine Developments - Health and Medicine | NBC News Clone | Ukraine Russia Conflict Updates - World News | NBC News Clone | Openai Chatgpt News - Technology and Innovation | NBC News Clone | 2024 Paris Games Highlights - Sports and Recreation | NBC News Clone | Extreme Weather Events - Weather and Climate | NBC News Clone | Hollywood Updates - Entertainment and Celebrity | NBC News Clone | Government Transparency - Investigations and Analysis | NBC News Clone | Community Stories - Local News and Communities | NBC News Clone