Technology News

Britannica, Merriam-Webster Sue OpenAI Over Copyright

A judge's gavel on a bench symbolizes the Britannica and Merriam-Webster lawsuit against OpenAI.

March 16, 2026 — Encyclopedia Britannica Inc. and its subsidiary Merriam-Webster have filed a federal lawsuit against OpenAI, accusing the artificial intelligence company of “massive copyright infringement.” The complaint alleges OpenAI used nearly 100,000 copyrighted online articles to train its large language models without permission or compensation.

Core Allegations in the Complaint

The lawsuit, filed in U.S. District Court, centers on three primary claims. First, Britannica alleges OpenAI illegally scraped its proprietary content to train models like GPT-4. The publisher retains copyright to its vast online database, which includes entries from its flagship encyclopedia and dictionary.

Second, the complaint argues OpenAI violates copyright when its models generate outputs containing “full or partial verbatim reproductions” of Britannica’s content. The suit also targets OpenAI’s use of retrieval-augmented generation (RAG), a technique where ChatGPT scans external databases for current information when answering queries.

“ChatGPT starves web publishers of revenue by generating responses that substitute, and directly compete with, content from publishers like Britannica,” the legal filing states. The publisher further contends that AI “hallucinations”—factually incorrect outputs falsely attributed to Britannica—damage its reputation and violate trademark law under the Lanham Act.

Broader Legal Landscape for AI Training

Britannica joins a growing list of content creators challenging AI companies in court. The New York Times, media conglomerate Ziff Davis, and more than a dozen U.S. and Canadian newspapers have filed similar suits against OpenAI. A separate Britannica lawsuit against AI firm Perplexity remains pending.

Legal precedent on whether using copyrighted material for AI training constitutes infringement remains unsettled. In a notable related case, AI company Anthropic successfully argued to U.S. District Judge William Alsup that using content as training data could be considered transformative fair use. However, Judge Alsup found Anthropic violated copyright law by illegally downloading millions of books without payment, leading to a $1.5 billion proposed settlement for affected authors.

This lack of clear legal framework places both publishers and AI developers in uncertain territory. The outcome of the Britannica case could establish important boundaries for the emerging industry.

Implications for Trust and Information Access

Beyond financial damages, Britannica’s lawsuit raises concerns about information integrity. The publisher argues that ChatGPT’s inaccuracies jeopardize “the public’s continued access to high-quality and trustworthy online information.” This positions the case as a conflict between rapid AI innovation and established editorial standards.

The complaint suggests AI-generated content could dilute reliable sources, making it harder for users to distinguish between verified information and algorithmic synthesis. This legal challenge arrives as AI tools become increasingly embedded in search engines and information services.

OpenAI did not respond to a request for comment prior to the lawsuit’s filing. The company has previously stated it respects the rights of content creators and believes its use of publicly available data falls within legal boundaries.

What Comes Next

The case will likely hinge on interpretations of fair use doctrine and how courts view the “transformative” nature of AI training. Legal experts anticipate a lengthy process that could eventually reach appellate courts. For now, the lawsuit signals intensified pressure on AI firms to establish clearer content licensing agreements with publishers. The resolution may determine whether AI companies need to fundamentally reshape how they acquire training data.

This article was produced with AI assistance and reviewed by our editorial team for accuracy and quality.

To Top