CondadosAI commited on
Commit
e6fdd15
·
verified ·
1 Parent(s): 6284133

docs: acaua mirror model card with code+weights provenance

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: acaua
4
+ pipeline_tag: keypoint-detection
5
+ tags:
6
+ - pose-estimation
7
+ - keypoint-detection
8
+ - vision
9
+ - acaua
10
+ - native-pytorch-port
11
+ - rtmpose
12
+ datasets:
13
+ - coco
14
+ - aic
15
+ ---
16
+
17
+ # RTMPose-tiny (COCO 17-keypoint) — acaua mirror (pure-PyTorch port)
18
+
19
+ This is a **pure-PyTorch port** of [RTMPose-tiny](https://github.com/open-mmlab/mmpose/tree/main/projects/rtmpose) hosted under `CondadosAI/` for use with the [acaua](https://github.com/CondadosAI/acaua) computer vision library.
20
+
21
+ The architecture has been re-implemented in pure PyTorch under `acaua.adapters.rtmpose` — no `mmcv`, no `mmengine`, no `mmpose`, no `trust_remote_code`. The weights in this mirror are converted from the upstream `.pth` checkpoint to safetensors with the acaua adapter's state_dict key naming, and load cleanly via `load_state_dict(strict=True)` into our nn.Module tree.
22
+
23
+ RTMPose is a **top-down** model: it consumes a person bounding box and predicts COCO 17-keypoint pose. The acaua adapter bundles [`CondadosAI/rtmdet_t_coco`](https://huggingface.co/CondadosAI/rtmdet_t_coco) as the person detector, giving you a single-call `predict(image)` API that returns boxes + keypoints together.
24
+
25
+ ## Provenance
26
+
27
+ | | |
28
+ |---|---|
29
+ | Upstream code (architecture) | [`open-mmlab/mmpose`](https://github.com/open-mmlab/mmpose) @ `759b39c13fea6ba094afc1fa932f51dc1b11cbf9` |
30
+ | Upstream code (backbone) | [`open-mmlab/mmdetection`](https://github.com/open-mmlab/mmdetection) @ `cfd5d3a985b0249de009b67d04f37263e11cdf3d` (CSPNeXt) |
31
+ | Upstream weights URL | `https://download.openmmlab.com/mmpose/v1/projects/rtmposev1/rtmpose-tiny_simcc-aic-coco_pt-aic-coco_420e-256x192-cfc8f33d_20230126.pth` |
32
+ | Upstream weights SHA256 | `e84eb5b9ee9432259bdd19d6a01156604ba27139ca6373ddb4ee7aa290d528e9` |
33
+ | Conversion script | [`scripts/convert_rtmpose.py`](https://github.com/CondadosAI/acaua/blob/main/scripts/convert_rtmpose.py) |
34
+ | Bundled detector | [`CondadosAI/rtmdet_t_coco`](https://huggingface.co/CondadosAI/rtmdet_t_coco) |
35
+ | Paper | Jiang et al., *"RTMPose: Real-Time Multi-Person Pose Estimation based on MMPose"*, arXiv:[2303.07399](https://arxiv.org/abs/2303.07399) |
36
+ | COCO val AP | 68.5 @ 256×192 (top-down, 17 keypoints) |
37
+ | Mirrored on | 2026-04-22 |
38
+ | Mirrored by | [CondadosAI/acaua](https://github.com/CondadosAI/acaua) |
39
+
40
+ ## Usage
41
+
42
+ ```python
43
+ import acaua
44
+ import supervision as sv
45
+
46
+ model = acaua.Model.from_pretrained("CondadosAI/rtmpose_t_coco")
47
+ result = model.predict("photo.jpg")
48
+
49
+ # `result` is a PoseResult: boxes (from RTMDet), keypoints (from RTMPose).
50
+ kp = result.to_supervision() # supervision.KeyPoints
51
+ sv.EdgeAnnotator(edges=model.skeleton).annotate(scene, kp)
52
+ ```
53
+
54
+ ## License and attribution
55
+
56
+ Redistributed under Apache-2.0, consistent with both the upstream code (mmpose / mmdetection, both Apache-2.0 by OpenMMLab) and the upstream weights declaration. The acaua adapter is a derivative work of the upstream PyTorch implementations — see [`NOTICE`](./NOTICE) for the required attribution chain (code AND weights).
57
+
58
+ ## Citation
59
+
60
+ ```bibtex
61
+ @misc{jiang2023rtmpose,
62
+ title={RTMPose: Real-Time Multi-Person Pose Estimation based on MMPose},
63
+ author={Tao Jiang and Peng Lu and Li Zhang and Ningsheng Ma and Rui Han and Chengqi Lyu and Yining Li and Kai Chen},
64
+ year={2023},
65
+ eprint={2303.07399},
66
+ archivePrefix={arXiv},
67
+ primaryClass={cs.CV}
68
+ }
69
+ ```