Consultancy Circle

Artificial Intelligence, Investing, Commerce and the Future of Work

BBC Contemplates Legal Action Against Perplexity AI’s Practices

In a significant development that could set critical legal precedents for the rapidly evolving AI landscape, the British Broadcasting Corporation (BBC) is threatening legal action against San Francisco-based AI company Perplexity AI over alleged copyright infringement. According to a report first published by Yahoo News on June 12, 2024, the BBC has demanded that Perplexity stop using its copyrighted content in AI-generated responses without authorization. The media giant is evaluating whether to seek an injunction and has issued a legal letter outlining its concerns. Perplexity AI, a growing challenger in the AI-assisted search market, is accused of “reproducing and communicating BBC content in breach of copyright,” sparking fresh concerns about the boundaries of fair use in generative AI.

What Sparked the Dispute Between BBC and Perplexity AI?

The dispute centers on how Perplexity AI, which utilizes large language models to generate answers to user queries, incorporates and displays content scrapped or derived from various sources—including media organizations like the BBC. According to the BBC, Perplexity’s tools often reproduce substantial passages from its articles verbatim, presenting them without clear attribution or direct user redirection to the original sources. This poses an existential question for content monetization models of traditional publishers in a world increasingly dominated by AI summary tools.

A BBC spokesperson stated that they are “concerned that content is being used in a way that undermines our ability to fund original journalism,” pointing to the public-service nature of the BBC’s operation funded through licensing fees. In its legal communication, the BBC indicated that use of its content in Perplexity’s AI responses may violate both UK and international copyright laws, including treaties such as the Berne Convention.

Legal and Ethical Implications for AI Content Generation

AI companies like Perplexity AI, OpenAI, and Google’s Gemini have all been navigating murky legal waters regarding dataset composition. The BBC case is unique in that it raises direct challenges around AI’s secondary use of journalism under the doctrine of fair use or fair dealing, differing by jurisdiction. These legal concepts, however, are historically applied in limited educational or commentary contexts—not commercial deployment at scale.

Recent legal precedents provide little clarity. Class-action lawsuits like the one against OpenAI by novelists Sarah Silverman and Paul Tremblay have yet to be resolved, while the New York Times’ landmark lawsuit filed in December 2023 against OpenAI and Microsoft is still pending. In light of the BBC’s latest stance, there is mounting pressure on governments and regulatory bodies to clarify whether LLM training and output mechanisms require licensing for content reproduction.

Competitive and Strategic Stakes for Perplexity AI

Perplexity AI, often dubbed the “AI-powered search engine challenger” to Google, has seen remarkable growth in 2024 and early 2025. As reported by VentureBeat, the company recently raised $73 million in a Series B funding round led by IVP, with backing from NEA, Elad Gil, and Nvidia CEO Jensen Huang. This capital injection brought its total valuation to over $1 billion. However, its strategic model heavily depends on aggregating and synthesizing web content to offer direct, conversational answers—hence the stakes in this legal confrontation soars beyond monetary damages.

Perplexity has positioned itself as a “truthful” alternative to hallucinating models, citing partnerships with trusted publishers and fact-checking pipelines. But these latest allegations could call that entire value proposition into question. If it’s determined that Perplexity is systematically misusing copyrighted text, the financial and reputational risks grow significantly—possibly threatening investor confidence and pipeline deals with publishers. Meanwhile, OpenAI, the reigning leader, has already entered licensing deals with Axel Springer and the Associated Press to forestall such disputes.

Industry Reactions and Regulatory Concerns

The BBC’s move adds to widening industry concern about the asymmetric power balance between AI labs and content creators. In mid-2024, the Federal Trade Commission (FTC) launched an inquiry into training data procurement practices by labs such as OpenAI and Anthropic, citing potential unfair competition and consumer deception issues (FTC Press Release, May 2024).

Moreover, in 2025, the European Union’s AI Act officially came into effect, containing specific provisions compelling high-risk AI systems to disclose and attribute their data sources transparently. Under Article 52(1) of the act, systems must ensure “traceability, transparency, and human oversight.” Meanwhile, the UK’s Competition and Markets Authority (CMA) issued a market study warning that AI developers must avoid anti-competitive behaviors, such as monopolizing access to data-rich private sources.

Ironically, the BBC’s own navigation of AI mirrors broader industry contradictions. The broadcaster itself utilizes AI tools for transcription, automated tagging, and headline generation. Nevertheless, the distinction between internal use and public-facing content synthesis appears to be a flashpoint.

The Cost Implications of Data Licensing and AI Deployment

This dispute resurfaces a silent reality of modern AI economics: access to up-to-date, high-quality datasets is no longer optional—it’s indispensable. The AI race is characterized by the exponential scaling of data acquisition expenditures. A July 2025 report from the McKinsey Global Institute estimates that data licensing costs for top-tier AI labs have risen by 140% since 2023, driven by increasing demand for legal clarity and exclusivity. If Perplexity were to lose or settle in favor of content licensing, the knock-on financial effect could radically reshape its budget.

Compare this with OpenAI and Google’s growing portfolio of content licensing deals, detailed in The Gradient’s January 2025 issue. OpenAI alone has reportedly spent over $250 million acquiring data rights through deals with Reuters, Red Ventures, and others—a strategic move to mitigate legal exposure for ChatGPT Enterprise and API clients.

Company Licensing Cost (Est. 2024–25) Major Content Deals
OpenAI $250M AP, Axel Springer, Reddit, Shutterstock
Google DeepMind $180M The Guardian, Stack Overflow, YouTube Creators Fund
Perplexity AI <$20M (est.) Limited; informal access

This table highlights the aggressive posture taken by market leaders to secure lawful AI operations, while startups like Perplexity risk exposure by shortcutting access methodologies. The BBC complaint may finally crystallize the calculus for emerging players: either pay for data or prepare for litigation.

What This Means for the Future of Journalism and AI

The broader implications of this case extend beyond Perplexity AI or even the BBC. As noted by the World Economic Forum in its 2025 Future of Work insights, journalism is increasingly being disintermediated. If readers receive summarized AI responses instead of clicking through to original articles, ad-based revenue systems collapse. This can undermine the sustainability of public interest journalism, investigative reporting, and global news delivery.

Solutions are emergent but not yet definitive. Some suggest “content labeling” frameworks where AI crawlers must adhere to robots.txt or receive API-facilitated access. Others posit that a collective bargaining mechanism, possibly modeled after music rights organizations like ASCAP or BMI, could centralize AI data licensing. Meanwhile, media outlets are considering AI partnerships or their own AI deployment strategies to proactively defend market share.

In the short term, the Perplexity–BBC standoff will likely become a litmus test for how AI startups negotiate legal risk and ethical responsibility during scale-up phases—especially when their core product is built on the intellectual property of others. As AI continues its steep adoption curve—with Gartner forecasting that 60% of knowledge workers will use AI daily by 2026—the demand for AI-native norms of content attribution and licensing will become unignorable.

The BBC’s legal response, while yet to materialize in formal court proceedings, may send ripples across Silicon Valley and beyond as regulatory interventions around AI content usage become more assertive and precise in the post-2024 era.

by Alphonse G

This article is based on or inspired by: Yahoo News, June 12, 2024

APA References:

  • VentureBeat. (2024). Perplexity AI raises $73 million to challenge ChatGPT and Google Search. Retrieved from https://venturebeat.com/ai/perplexity-ai-raises-73-million-to-take-on-chatgpt-and-google-in-web-search/
  • FTC. (2024). FTC launches inquiry into training data used by AI labs. Retrieved from https://www.ftc.gov/news-events/news/press-releases
  • McKinsey Global Institute. (2025). AI scaling economics in 2025: The new battleground is data. Retrieved from https://www.mckinsey.com/mgi
  • The Gradient. (2025). January Issue: Licensing in the Age of AI. Retrieved from https://thegradient.pub/
  • World Economic Forum. (2025). AI and the Future of Journalism. Retrieved from https://www.weforum.org/focus/future-of-work
  • Yahoo News. (2024). BBC threatens injunction against Perplexity AI. Retrieved from https://www.yahoo.com/news/bbc-threatens-injunction-against-perplexity-120753690.html
  • MIT Technology Review. (2025). The path to ethical AI: Content licensing strategies. Retrieved from https://www.technologyreview.com/topic/artificial-intelligence/
  • OpenAI Blog. (2024). Partnering for responsible AI. Retrieved from https://openai.com/blog/
  • NVIDIA Blog. (2024). Beyond GPUs: Building data-ready AI infrastructure. Retrieved from https://blogs.nvidia.com/
  • DeepMind Blog. (2025). Training responsibly in a multi-stakeholder ecosystem. Retrieved from https://www.deepmind.com/blog

Note that some references may no longer be available at the time of your reading due to page moves or expirations of source articles.