Skip to main content
    Skip to main contentSkip to navigationSkip to footer
    Artificial Intelligence
    (Adversarielle Robustheit)

    Adversarial Robustness

    Also known as:
    Adversarial Defense
    Robust ML
    Attack Resilience
    Updated: 2/11/2026

    The ability of an ML model to maintain correct predictions even when inputs are deliberately manipulated.

    Quick Summary

    Adversarial robustness makes ML models resilient against deliberate input manipulations – essential for safe AI in production.

    Explanation

    Adversarial robustness is achieved through adversarial training, certified defenses, input preprocessing, or randomized smoothing. Trade-offs between robustness and accuracy are unavoidable.

    Marketing Relevance

    For marketing AI in production (content moderation, fraud detection), adversarial robustness is critical for trust and security.

    Example

    A spam filter is hardened through adversarial training against Unicode tricks and homoglyph attacks.

    Common Pitfalls

    Robustness against one attack doesn't protect against all attacks. Adversarial training is compute-intensive and can reduce accuracy.

    Origin & History

    Madry et al. (2018) established PGD-based adversarial training as the gold standard. Certified defenses (Randomized Smoothing, Cohen et al. 2019) provided formal guarantees. RobustBench standardized benchmarking from 2021.

    Comparisons & Differences

    Adversarial Robustness vs. Adversarial Attacks

    Adversarial attacks are the attack methods; adversarial robustness is the defense capability against them.

    Adversarial Robustness vs. Robustness Testing

    Robustness testing evaluates general reliability; adversarial robustness specifically focuses on protection against targeted attacks.

    Related Services

    Related Terms

    👋Questions? Chat with us!