Model Extraction
Attacks that attempt to reconstruct or clone a proprietary ML model through systematic queries.
Model Extraction clones proprietary AI models through systematic API queries. Billions in R&D can be stolen. Rate limiting and query monitoring are essential.
Explanation
Attacker sends many queries to API, collects input-output pairs, trains "surrogate model". Works with MLaaS, can steal billions in R&D investment. Defenses: Rate limiting, query monitoring, output perturbation.
Marketing Relevance
Anyone offering custom AI models via APIs risks model extraction. Competitors could steal proprietary insights.
Example
A competitor sends 1 million queries to a product recommendation API, trains their own model – saves years of development and data collection.
Common Pitfalls
Hard to distinguish from legitimate usage. Rate limiting can hinder real customers. Legal situation unclear.
Origin & History
Tramèr et al. demonstrated model extraction against cloud ML APIs in 2016. With the LLM era, it became relevant for API services like OpenAI. API access costs make attacks more expensive, but not impossible.
Comparisons & Differences
Model Extraction vs. Data Poisoning
Model Extraction wants to steal the model; Data Poisoning wants to manipulate model behavior.
Model Extraction vs. Prompt Leaking
Prompt Leaking extracts system prompts; Model Extraction wants to clone entire model knowledge.