HB Ad Slot
HB Mobile Ad Slot
Changes Proposed to the Federal Rules of Evidence to Address AI Usage
Friday, November 15, 2024

Highlights

  • Proposed amendments to federal rules of evidence address authentication of evidence generated by artificial intelligence (AI)
  • If an AI output is offered as evidence, and that type of evidence would traditionally have required an expert witness’ testimony, it would be subject to a Rule 702 analysis under the new rule
  • To authenticate evidence generated by AI, the proponent of the evidence would need to provide sufficient detail on training data, AI program used, and reliability of the AI outputs

If the opponent of a piece of evidence can “reasonably” demonstrate it has been altered by AI, the evidence would only be admissible under the new rule if it's “more likely than not authentic”

The U.S. Courts Advisory Committee on the Federal Rules of Evidence has offered proposed amendments to rule changes to address the use of artificial intelligence (AI) in litigation. The proposed amendments would expand upon Rule 901 (Authenticating or Identifying Evidence) and would create a new rule – Rule 707, “Machine-generated Evidence.”

The proposed amendments are included in the Agenda Book for the Committee’s November meeting at pages 269-271.

Changes to Rule 901

Rule 901(a) provides that “to satisfy the requirement of authenticating or identifying an item of evidence, the proponent must produce evidence sufficient to support a finding that the item is what the proponent claims it is.” Subsection (b) then provides specific examples of the types of evidence that satisfy the requirements of section (a).

The proposed amendments would add language to the list of examples that describes what is needed to demonstrate the authenticity of evidence that is “generated by artificial intelligence.” Under the amended rule, the proponent of such evidence would need to produce evidence that, among others “(i) describes the training data and software or program that was used; and (ii) shows that they produced reliable results in this instance.”

Additionally, the proposed amendment adds a new section – subsection (c) – to directly address “deepfakes” and the burden for advancing or opposing evidence that is suspected of being “altered or fabricated, in whole or in part, by artificial intelligence.” This section would include a two-step test, with a shifting burden, when the opponent of a piece of evidence alleges alteration or fabrication by artificial intelligence.

Initially, the opponent of the evidence must demonstrate “to the court that a jury reasonably could find” that the evidence has been altered. Upon such a showing, the burden then shifts to the proponent, and the evidence is “admissible only if the proponent demonstrates to the court that it is more likely than not authentic.”

New Rule 707

The proposed amendments also seek to subject AI outputs to the same standard used to assess the admissibility of expert witness testimony, namely Rule 702. Under proposed Rule 707, “[w]here the output of a process or system would be subject to Rule 702 if testified to by a human witness, the court must find that the output satisfies the requirements of Rule 702 (a)-(d).

For example, a damages expert in a business dispute would normally look at factors that relate to the business' performance and would apply those figures through a formula. The expert would then testify as to the reasonableness and reliability of the methodology. However, AI cannot testify for itself as to how it arrives at its output.

Therefore, if AI is used to calculate the final damages amount, then the proponent would need to demonstrate: a) that the output would help the trier of fact, b) sufficient facts or data were used as the inputs for the AI program, c) the AI program used reliable principles and methods, and d) that the output reflects a reliable application of the principles and methods to the inputs.

To demonstrate that these requirements are met, the committee noted courts would consider what inputs are used, ensure that the opponent has sufficient access to the AI program to evaluate its functioning, and consider whether the process has been validated in sufficiently similar circumstances.

There are several purposes for subjecting AI-outputs to the standard of reliability applied to expert witnesses: to prevent against function creep, analytical error, inaccuracy or bias, and lack of interpretability.

The proposed amendment specifically exempts “basic scientific instruments or routinely relied upon commercial software, which the committee noted would include outputs of non-AI tools such as “a mercury-based thermometer, battery-operated digital thermometer, or automated averaging of data in a spreadsheet.”

Takeaways

As the use of AI tools in litigation expands, courts and rules committees are addressing how best to manage the use of AI generated information as evidence. These proposed amendments, which are still being considered, are designed to ensure the authenticity and reliability of evidence presented to the trier of fact. Legal practitioners should aim to stay up to date on all evidentiary rules that may impact when and how they may use certain AI outputs to prove their cases.

HTML Embed Code
HB Ad Slot
HB Ad Slot
HB Mobile Ad Slot
HB Ad Slot
HB Mobile Ad Slot
 
NLR Logo
We collaborate with the world's leading lawyers to deliver news tailored for you. Sign Up to receive our free e-Newsbulletins

 

Sign Up for e-NewsBulletins