Skip to main content
    Skip to main contentSkip to navigationSkip to footer
    Artificial Intelligence

    Pose Estimation

    Also known as:
    Human Pose Estimation
    Body Pose Detection
    Skeleton Detection
    Keypoint Detection
    Updated: 2/10/2026

    Detection and localization of body joints and skeleton keypoints in images or videos.

    Quick Summary

    Pose estimation detects body joints and skeletons in images – foundation for fitness apps, sports analysis, AR/VR, and gesture recognition.

    Explanation

    Pose estimation typically detects 17-25 keypoints (eyes, shoulders, elbows, knees, etc.) and connects them into a skeleton. Top-down approaches first detect people then poses; bottom-up detects all keypoints simultaneously.

    Marketing Relevance

    Pose estimation is central to fitness apps, AR/VR, sports analysis, physiotherapy, and gesture recognition.

    Example

    A fitness app detects body posture during exercise and provides real-time feedback on correct form.

    Common Pitfalls

    Occlusions by other people or objects. Weaknesses with unusual poses. High compute for multi-person real-time.

    Origin & History

    DeepPose (Google, 2014) brought deep learning to pose estimation. OpenPose (CMU, 2017) enabled multi-person real-time detection. MediaPipe (Google, 2019) made pose estimation available on mobile. ViTPose (2022) uses Vision Transformers.

    Comparisons & Differences

    Pose Estimation vs. Object Detection

    Object detection finds bounding boxes. Pose estimation finds finer skeleton keypoints within detected people.

    Pose Estimation vs. Action Recognition

    Pose estimation detects body posture in a frame. Action recognition classifies activities across time sequences.

    Related Services

    Related Terms

    Computer VisionObject DetectionAction RecognitionMediaPipeOpenPose
    👋Questions? Chat with us!