Back to Articles

Google Says People Are Copying Its AI, Much Like It Scraped Theirs

Google Says People Are Copying Its AI, Much Like It Scraped Theirs

Key Takeaways

  1. 1Google claims its Gemini AI model is under attack from "commercially motivated" actors attempting to clone it.
  2. 2The attackers are allegedly using up to 100,000 queries to "extract" the model’s logic, a process Google calls a “distillation attack.”
  3. 3Google defends its intellectual property while facing criticism for using vast amounts of data without compensating creators to train its AI.
  4. 4The incident underscores the challenges of protecting AI models that are offered as services via Application Programming Interfaces (APIs).

Google is accusing others of stealing its AI models using "distillation attacks," a technique to reverse engineer the underlying model via excessive queries. The accusation lands with a thud given Google's own history of scraping data to train its AI, highlighting a growing tension around intellectual property in the AI space.

The Pot Calling the AI Kettle Black

Google says it has detected attempts to steal its Gemini AI model via "distillation attacks." According to Google, these attacks involve sending a massive number of prompts—up to 100,000—to the model in an attempt to replicate its reasoning abilities. This is akin to reverse-engineering the model through sheer persistence.

Distillation Attacks Explained

A distillation attack (or model extraction attack) occurs when bad actors try to extract the functionality of a large, proprietary model by querying it repeatedly and analyzing the responses. Think of it as trying to figure out how a complicated machine works by only looking at what it produces.

Google characterized these actions as "intellectual property theft" and a violation of its terms of service. The company stated that the attacks targeted Gemini’s ability to reason across multiple languages.

The Hypocrisy Angle

Google's claims of intellectual property theft have been met with skepticism, due to its practice of scraping vast amounts of data from the internet without permission to train its AI models. This has led to several copyright infringement lawsuits.

The irony isn't lost on observers. Google is now claiming foul over similar behavior, suggesting a double standard when it comes to using data for AI development.

AI Model Vulnerability

Google’s complaint highlights a broader vulnerability in the AI industry. As Large Language Models (LLMs) become more powerful, protecting them from unauthorized duplication becomes increasingly difficult. This challenge is particularly acute when these models are offered as services via APIs.

"For many AI technologies where LLMs are offered as services, this approach is no longer required; actors can use legitimate API access to attempt to 'clone' select AI model capabilities," Google's report states.

A Race to Monetize

AI companies are under pressure to monetize their technologies, leading to a variety of revenue models, from subscriptions to advertising. Protecting intellectual property is key to these strategies.

News of Google's troubles comes as Google brings agentic shopping to AI search, letting US shoppers buy items from Etsy and Wayfair in AI Mode in Search as well as the Gemini app. This shows Google's commitment to integrate AI in the e-commerce experience.

What's Next

Expect increased legal and technical efforts to protect AI models from extraction. Companies will likely invest in more robust monitoring and defense mechanisms to detect and mitigate distillation attacks. We may also see more stringent terms of service and API usage policies.

Why It Matters

    • IP Protection: This incident emphasizes the need for stronger intellectual property protection for AI models.
    • Ethical Concerns: It raises ethical questions about data usage and the responsibilities of AI companies.
    • Model Security: It underscores the vulnerability of AI models accessible through APIs.
    • Innovation Impact: The ability to protect AI models will directly impact ongoing investments in AI research and development.
    • Industry Debate: As AI interfaces feel more human, marketing becomes the first point of ethical exposure, and AI companies may face increasing scrutiny.


Source: futurism.com

Disclosure: This article is for informational purposes only.

FAQ

AI distillation attacks, also known as model extraction attacks, involve repeatedly querying an AI model to reverse engineer its underlying logic and functionality. Attackers send a massive number of prompts to the model, then analyze the responses to replicate the AI's reasoning abilities. Google characterizes these attacks as intellectual property theft when used against models like Gemini.

Google's accusation of AI copying is controversial because the company itself has a history of scraping vast amounts of data from the internet to train its AI models, often without compensating the original creators. This practice has led to copyright infringement lawsuits and accusations of hypocrisy, as Google now claims foul over similar data usage practices.

AI companies are investing in more robust monitoring and defense mechanisms to detect and mitigate distillation attacks. They are also implementing more stringent terms of service and API usage policies to protect their intellectual property. Stronger legal and technical efforts are expected to protect AI models from unauthorized extraction.

Google's Gemini is a large language model (LLM) that is offered as a service via APIs. Google claims that commercially motivated actors are attempting to clone it using up to 100,000 queries to extract the model’s logic, which Google calls a “distillation attack.”

Protecting intellectual property is key to the monetization strategies of AI companies. As Large Language Models (LLMs) become more powerful, protecting them from unauthorized duplication becomes increasingly difficult, especially when offered as services via APIs. This incident emphasizes the need for stronger intellectual property protection for AI models.

Newsletter

Stay informed without the noise.

Daily AI updates for builders. No clickbait. Just what matters.