Back to news

Artificial Intelligence Hub

May 17, 2025

Litigation Risks from AI Use: A Growing Concern

Companies face rising AI, data and compliance risks as regulations expand. Managing privacy, cybersecurity and complex system integration is increasingly critical.

Companies are encountering an increasing number of complexities when it comes to managing vast data flows and the intricate integration of multiple systems. The evolution of new technologies, especially AI, adds layers of difficulty in ensuring that all systems are interconnected and compliant with growing regulatory demands. Companies are under intense scrutiny, not just to ensure smooth operations but also to navigate an ever-expanding set of regulatory frameworks governing data privacy, cybersecurity and AI use.

Compounding these challenges is the inherent opacity of AI technology itself. AI systems operate in ways that are often difficult for humans to understand fully, making compliance even more difficult. Traditional governance and compliance frameworks, developed for a slower-paced and more transparent digital ecosystem, struggle to address the unique risks posed by AI effectively. With its scale, complexity and ability to evolve quickly, AI demands a more sophisticated, dynamic and transparent approach to governance - one that existing regulatory models may not fully capture.

Increasing Privacy and Cyber Risk Exposure

Privacy litigation now extends beyond traditional breaches, with significant class action settlements, such as those resulting from pixel-tracking, in the US. In the EU, regulators focus on technical violations of GDPR, especially concerning AI implementations and third-party data sharing. Meanwhile, cyber risks are rising, with organisations facing regulatory enforcement from NIS2 in the UK, compromised supply chains and securities litigation linked to undisclosed digital vulnerabilities. New generative AI capabilities, such as model poisoning and prompt injection, add complexity, creating new areas requiring proactive cybersecurity measures.

Who Owns the Data and the Output?

AI challenges traditional IP frameworks, as systems often consume and transform protected content on an unprecedented scale. This raises legal risks for downstream users relying on commercial models trained on contested data. Legal disputes are intensifying as generative models are trained on large amounts of copyrighted content, leading to litigation in various jurisdictions. Visual AI models, in particular, are at the heart of disputes involving traditional publishers, content creators and AI firms.

Trustworthiness of Technology

AI systems can reflect or amplify biases in training data or design, leading to legal exposure under discrimination and human rights laws. This is particularly evident in sectors like employment, finance and law enforcement. The <span class="news-text_italic-underline">EU AI Act</span> permits monitoring for bias in high-risk AI but requires compliance with GDPR guidelines.

Explainability and Transparency

For industries such as healthcare, lending and admissions, explainability in AI is essential, yet many systems remain “black boxes”.

A Complex and Evolving Regulatory Landscape

AI regulation spans multiple areas: data protection, safety, competition, cybersecurity and sector-specific regimes. The intersection of these regulatory frameworks creates complex, overlapping obligations for AI systems, both in development and deployment.

  • <span class="news-text_medium">Data Protection:</span> GDPR and equivalent global laws govern data use, automated decision-making and profiling.
  • <span class="news-text_medium">AI-Specific:</span> The <span class="news-text_italic-underline">EU AI Act</span> and the UK’s sector-led approach establish principles-based obligations.
  • <span class="news-text_medium">Online Safety:</span> UK and EU regulations impose duties for transparency and accountability on algorithmic systems.
  • <span class="news-text_medium">Product Liability:</span> EU and UK frameworks cover AI-enabled products and autonomous systems.
  • <span class="news-text_medium">Cybersecurity:</span> NIS2 requires reporting, testing and resilience for AI systems in critical sectors.
  • <span class="news-text_medium">Competition:</span> <span class="news-text_italic-underline">Digital Markets Act</span> and UK reforms address algorithmic collusion and data access.

As technology evolves faster than laws can keep up, a gap emerges between regulatory intent and enforcement. This disconnect complicates how organisations respond to incidents, requiring experts with hands-on experience in AI evidence handling and knowledge of regulatory changes.

Reputation Risks from AI Litigation

The impact of AI litigation on an organisation’s reputation can be swift and severe, driven by rapid technological change and intense media scrutiny. Managing stakeholder perceptions and understanding the broader societal implications of AI regulation is critical. Changes in regulation can affect investment, innovation and economic stability.

Reputation Risks Breakdown:

  • <span class="news-text_medium">Business-to-Business:</span> IP litigation between rights holders and AI platforms can harm reputations through commercial disputes.
  • <span class="news-text_medium">Business-to-Consumer:</span> Discrimination claims can create "David vs Goliath" battles in the court of public opinion, especially if AI fails to meet user expectations.
  • <span class="news-text_medium">Business-to-Government/Regulator:</span> Breaches of AI-related laws can damage public goodwill and require substantial political capital for reform.

What’s Next?

The intersection of AI regulation with data protection, product liability, competition and cybersecurity laws creates complex obligations for AI systems. While AI’s potential for increased productivity is appealing, organisations must understand their legal, operational and strategic risks of AI in litigation. Firms addressing these risks early will not only build safeguards but will also shape industry standards, enhance resilience and gain a competitive edge.

Address
London:
2 Eaton Gate
London SW1W 9BJ
New York:
295 Madison Ave 12th Floor
New York City, NY 10017
Paris:
56 Avenue Kléber,
75116 Paris
Moscow:
Molodegnaya st., build. 5
119296, Moscow
BELGRAVIA LAW LIMITED is registered with the Solicitors Regulation Authority with SRA number 8004056 and is a limited company registered in England & Wales with company number 14815978. The firm’s registered office is at 2 Eaton Gate, Belgravia, London SW1W 9BJ.

‘Belgravia Law’ (c) 2025. All rights reserved.
By clicking “Accept”, you agree to the storing of cookies on your device to enhance site navigation, analyse site usage, and assist in our marketing efforts. View our Privacy Policy and Cookie Policy for more information.