docs: acaua mirror model card with code+weights provenance
Browse files
README.md
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
library_name: acaua
|
| 4 |
+
pipeline_tag: keypoint-detection
|
| 5 |
+
tags:
|
| 6 |
+
- pose-estimation
|
| 7 |
+
- keypoint-detection
|
| 8 |
+
- vision
|
| 9 |
+
- acaua
|
| 10 |
+
- native-pytorch-port
|
| 11 |
+
- rtmpose
|
| 12 |
+
datasets:
|
| 13 |
+
- coco
|
| 14 |
+
- aic
|
| 15 |
+
---
|
| 16 |
+
|
| 17 |
+
# RTMPose-tiny (COCO 17-keypoint) — acaua mirror (pure-PyTorch port)
|
| 18 |
+
|
| 19 |
+
This is a **pure-PyTorch port** of [RTMPose-tiny](https://github.com/open-mmlab/mmpose/tree/main/projects/rtmpose) hosted under `CondadosAI/` for use with the [acaua](https://github.com/CondadosAI/acaua) computer vision library.
|
| 20 |
+
|
| 21 |
+
The architecture has been re-implemented in pure PyTorch under `acaua.adapters.rtmpose` — no `mmcv`, no `mmengine`, no `mmpose`, no `trust_remote_code`. The weights in this mirror are converted from the upstream `.pth` checkpoint to safetensors with the acaua adapter's state_dict key naming, and load cleanly via `load_state_dict(strict=True)` into our nn.Module tree.
|
| 22 |
+
|
| 23 |
+
RTMPose is a **top-down** model: it consumes a person bounding box and predicts COCO 17-keypoint pose. The acaua adapter bundles [`CondadosAI/rtmdet_t_coco`](https://huggingface.co/CondadosAI/rtmdet_t_coco) as the person detector, giving you a single-call `predict(image)` API that returns boxes + keypoints together.
|
| 24 |
+
|
| 25 |
+
## Provenance
|
| 26 |
+
|
| 27 |
+
| | |
|
| 28 |
+
|---|---|
|
| 29 |
+
| Upstream code (architecture) | [`open-mmlab/mmpose`](https://github.com/open-mmlab/mmpose) @ `759b39c13fea6ba094afc1fa932f51dc1b11cbf9` |
|
| 30 |
+
| Upstream code (backbone) | [`open-mmlab/mmdetection`](https://github.com/open-mmlab/mmdetection) @ `cfd5d3a985b0249de009b67d04f37263e11cdf3d` (CSPNeXt) |
|
| 31 |
+
| Upstream weights URL | `https://download.openmmlab.com/mmpose/v1/projects/rtmposev1/rtmpose-tiny_simcc-aic-coco_pt-aic-coco_420e-256x192-cfc8f33d_20230126.pth` |
|
| 32 |
+
| Upstream weights SHA256 | `e84eb5b9ee9432259bdd19d6a01156604ba27139ca6373ddb4ee7aa290d528e9` |
|
| 33 |
+
| Conversion script | [`scripts/convert_rtmpose.py`](https://github.com/CondadosAI/acaua/blob/main/scripts/convert_rtmpose.py) |
|
| 34 |
+
| Bundled detector | [`CondadosAI/rtmdet_t_coco`](https://huggingface.co/CondadosAI/rtmdet_t_coco) |
|
| 35 |
+
| Paper | Jiang et al., *"RTMPose: Real-Time Multi-Person Pose Estimation based on MMPose"*, arXiv:[2303.07399](https://arxiv.org/abs/2303.07399) |
|
| 36 |
+
| COCO val AP | 68.5 @ 256×192 (top-down, 17 keypoints) |
|
| 37 |
+
| Mirrored on | 2026-04-22 |
|
| 38 |
+
| Mirrored by | [CondadosAI/acaua](https://github.com/CondadosAI/acaua) |
|
| 39 |
+
|
| 40 |
+
## Usage
|
| 41 |
+
|
| 42 |
+
```python
|
| 43 |
+
import acaua
|
| 44 |
+
import supervision as sv
|
| 45 |
+
|
| 46 |
+
model = acaua.Model.from_pretrained("CondadosAI/rtmpose_t_coco")
|
| 47 |
+
result = model.predict("photo.jpg")
|
| 48 |
+
|
| 49 |
+
# `result` is a PoseResult: boxes (from RTMDet), keypoints (from RTMPose).
|
| 50 |
+
kp = result.to_supervision() # supervision.KeyPoints
|
| 51 |
+
sv.EdgeAnnotator(edges=model.skeleton).annotate(scene, kp)
|
| 52 |
+
```
|
| 53 |
+
|
| 54 |
+
## License and attribution
|
| 55 |
+
|
| 56 |
+
Redistributed under Apache-2.0, consistent with both the upstream code (mmpose / mmdetection, both Apache-2.0 by OpenMMLab) and the upstream weights declaration. The acaua adapter is a derivative work of the upstream PyTorch implementations — see [`NOTICE`](./NOTICE) for the required attribution chain (code AND weights).
|
| 57 |
+
|
| 58 |
+
## Citation
|
| 59 |
+
|
| 60 |
+
```bibtex
|
| 61 |
+
@misc{jiang2023rtmpose,
|
| 62 |
+
title={RTMPose: Real-Time Multi-Person Pose Estimation based on MMPose},
|
| 63 |
+
author={Tao Jiang and Peng Lu and Li Zhang and Ningsheng Ma and Rui Han and Chengqi Lyu and Yining Li and Kai Chen},
|
| 64 |
+
year={2023},
|
| 65 |
+
eprint={2303.07399},
|
| 66 |
+
archivePrefix={arXiv},
|
| 67 |
+
primaryClass={cs.CV}
|
| 68 |
+
}
|
| 69 |
+
```
|