BBoxMaskPose v2 (BMPv2)
- The paper introduces BMPv2, a multi-stage framework that integrates probabilistic, mask-conditioned 2D pose estimation with iterative mask refinement and a mutual conditioning loop.
- It achieves state-of-the-art performance on COCO and OCHuman benchmarks, surpassing 50% AP on challenging datasets and boosting 3D pose quality.
- The approach demonstrates that precise 2D keypoint inference and occlusion modeling are key to improving multi-person 3D pose estimation in crowded scenes.
BBoxMaskPose v2 (BMPv2) is a multi-stage pose estimation framework that advances multi-person 2D and 3D human pose estimation, particularly in scenes with significant person overlap. BMPv2 integrates a probabilistic, mask-conditioned 2D pose estimation head (PMPose), an enhanced mask refinement module based on Segment Anything Model (SAM), and a tightly coupled mutual conditioning loop. BMPv2 sets new state-of-the-art (SOTA) performance on both COCO and OCHuman benchmarks, and for the first time surpasses 50 average precision (AP) on OCHuman. The method also establishes that improvements in 2D pose quality produce commensurate gains in subsequent 3D pose estimation, as verified on the newly introduced OCHuman-Pose dataset (Purkrabek et al., 21 Jan 2026).
1. Model Architecture and Mutual Conditioning
The BMPv2 pipeline operates as a modular sequence consisting of three core components: (A) object detection, (B) 2D pose estimation with PMPose, and (C) mask refinement using SAM-pose2seg.
PMPose Head:
PMPose is a top-down, probabilistic, mask-conditioned 2D pose estimator. Inputs are image crops centered on detected boxes and binary masks . The backbone leverages ViT-S/B/L/H variants (following the ViTPose paradigm), producing feature maps . Mask conditioning is realized through element-wise addition or multiplication of a learned embedding of to . For each of the keypoints, the PMPose head outputs:
- Heatmap (discrete probability over pixel locations)
- Presence probability
- Visibility probability
- Predicted expected OKS
The full instance-level probabilistic output is: with .
Mutual Conditioning Loop:
BMPv2 sequentially applies:
- Detector: Ingests image and previous background mask, outputs a new box.
- PMPose: Ingests image, current box, and previous mask, outputs keypoints and probabilities.
- SAM-pose2seg: Refines the mask, using pose-guided prompts derived from PMPose outputs.
BMPv2+ adds an additional pass of pose and mask refinement for further accuracy gains. Conditioning each module on others’ predictions iteratively "un-collapses" overlapping instances, resolving ambiguities in crowded settings.
SAM-pose2seg Mask Refinement:
SAM-pose2seg is a specialized fine-tuning of SAM v2.1, with the image encoder frozen and decoder trained only on the “human” class. Mask prompts are generated using PMPose: the prompt is set to the joint of highest visibility probability; further prompts target joints within current error regions.
The mask segmentation loss (per mask) is:
2. Training Objectives and Loss Formulation
PMPose Loss:
Given ground-truth heatmaps , presence , visibility , and OKS , the PMPose learning objective is:
Mask conditioning is injected via: where and are learnable projections.
3D Regression Losses:
BMPv2’s outputs can be directly used as prompts for 3D mesh recovery models (e.g., SAM-3D-Body), predicting 3D joints and shape codes :
3. Mask-Conditioned Prompting for 3D Pose Recovery
BMPv2 establishes a direct pipeline for using 2D mask- and pose-conditioned descriptors to drive 3D pose and shape estimation. The 3D module concatenates flattened full-image features , boxes , masks , and all heatmaps into a single vector:
Inclusion of mask prompts facilitates precise occlusion reasoning and boundary localization absent from box-only inputs. Reported experimental results show that on OCHuman-Pose, reprojection AP increases from approximately 75 (box prompt alone) to 85 when both mask and pose prompts are used. This suggests that mask conditioning is a principal driver of improvements in challenging crowd scenarios.
4. Empirical Performance on COCO and OCHuman
BMPv2 demonstrates substantial advances over prior art on standard pose estimation benchmarks. Key results, including comparisons to relevant baselines, are summarized below.
| Model | OCHuman-test AP | OCHuman-Pose-test AP | COCO-val AP |
|---|---|---|---|
| ViTPose-B | 37.5 | 67.8 | 76.4 |
| MaskPose-B | 46.6 | 78.0 | 76.8 |
| PMPose-B | 48.2 | 80.0 | 76.9 |
| BMPv2 2 | 51.5 | 83.4 | 78.8 |
| BMPv2+ 2 | 55.8 | 86.8 | 78.1 |
BMPv2 is the first method to exceed 50% AP on OCHuman-test. BMPv2+ achieves up to 4 points higher AP in exchange for additional compute, underscoring the effectiveness of extended mutual conditioning.
5. Analysis with OCHuman-Pose and Error Attribution
The OCHuman-Pose dataset provides a more comprehensive multi-person testbed by adding annotations covering over 50% more subjects. It closes the typical gap between COCO-style AP with detected boxes and ground-truth box-based AP, ensuring that benchmarking more accurately reflects method differences.
Analysis reveals that using RTMDet-L + ViTPose* as a baseline in 2D yields 3D reprojection AP near 76. When masks are refined via BMPv2 and supplied to SAM-3D-Body, 3D reprojection AP increases to around 85. Thus, in crowded settings, the BMP loop can mitigate detection errors from merged or missed boxes, shifting the performance bottleneck in 3D pose estimation to the quality of 2D pose keypoint association.
6. Significance and Implications
BMPv2 establishes that integrating probabilistic, mask-conditioned 2D pose heads with iterative mask refinement and tightly coupled module conditioning is highly effective for multi-person human pose estimation in crowded scenes. Improvements in 2D keypoint accuracy, rather than bounding box detection, are now the limiting factor for multi-person 3D pose estimation in complex images. A plausible implication is that future progress in the field will depend more on advancing 2D keypoint inference, association, and occlusion modeling than on purely detection-based approaches.
The released code, models, and the OCHuman-Pose dataset facilitate further research and deployment in both 2D and 3D multi-person understanding tasks (Purkrabek et al., 21 Jan 2026).