Skip to main content
    Skip to main contentSkip to navigationSkip to footer
    Artificial Intelligence

    Model Extraction

    Also known as:
    Model Stealing
    Model Theft
    Model Cloning
    Knowledge Extraction
    Updated: 2/9/2026

    Attacks that attempt to reconstruct or clone a proprietary ML model through systematic queries.

    Quick Summary

    Model Extraction clones proprietary AI models through systematic API queries. Billions in R&D can be stolen. Rate limiting and query monitoring are essential.

    Explanation

    Attacker sends many queries to API, collects input-output pairs, trains "surrogate model". Works with MLaaS, can steal billions in R&D investment. Defenses: Rate limiting, query monitoring, output perturbation.

    Marketing Relevance

    Anyone offering custom AI models via APIs risks model extraction. Competitors could steal proprietary insights.

    Example

    A competitor sends 1 million queries to a product recommendation API, trains their own model – saves years of development and data collection.

    Common Pitfalls

    Hard to distinguish from legitimate usage. Rate limiting can hinder real customers. Legal situation unclear.

    Origin & History

    Tramèr et al. demonstrated model extraction against cloud ML APIs in 2016. With the LLM era, it became relevant for API services like OpenAI. API access costs make attacks more expensive, but not impossible.

    Comparisons & Differences

    Model Extraction vs. Data Poisoning

    Model Extraction wants to steal the model; Data Poisoning wants to manipulate model behavior.

    Model Extraction vs. Prompt Leaking

    Prompt Leaking extracts system prompts; Model Extraction wants to clone entire model knowledge.

    Related Services

    Related Terms

    ai-securityapi-securityintellectual-propertyMLOpsAdversarial Attacks
    👋Questions? Chat with us!