Update README.md
Browse files
README.md
CHANGED
@@ -32,16 +32,20 @@ download_huggingface_model("superanimal_quadruped", model_dir)
|
|
32 |
## Intended Use
|
33 |
• Intended to be used for pose estimation of quadruped images taken from side-view. The model serves a better starting
|
34 |
point than ImageNet weights in downstream datasets such as AP-10K.
|
|
|
35 |
• Intended for academic and research professionals working in fields related to animal behavior, such as neuroscience
|
36 |
and ecology.
|
|
|
37 |
• Not suitable as a zeros-shot model for applications that require high keypiont precision, but can be fine-tuned with
|
38 |
minimal data to reach human-level accuracy. Also not suitable for videos that look dramatically different from those
|
39 |
we show in the paper.
|
40 |
-
|
|
|
|
|
41 |
• Based on the known robustness issues of neural networks, the relevant factors include the lighting, contrast and
|
42 |
resolution of the video frames. The present of objects might also cause false detections and erroneous keypoints.
|
43 |
When two or more animals are extremely close, it could cause the top-down detectors to only detect only one animal,
|
44 |
-
if used without further fine-tuning or with a method such as BUCTD (
|
45 |
|
46 |
## Metrics
|
47 |
• Mean Average Precision (mAP)
|
@@ -78,12 +82,6 @@ Here is an image with the keypoint guide:
|
|
78 |
<img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690988780004-AG00N6OU1R21MZ0AU9RE/modelcard-SAQ.png?format=1500w" width="95%">
|
79 |
</p>
|
80 |
|
81 |
-
Please note that each dataset was labeled by separate labs \& separate individuals, therefore while we map names
|
82 |
-
to a unified pose vocabulary (found here: https://github.com/AdaptiveMotorControlLab/modelzoo-figures), there will be annotator bias in keypoint placement (See the Supplementary Note on annotator bias).
|
83 |
-
You will also note the dataset is highly diverse across species, but collectively has more representation of domesticated animals like dogs, cats, horses, and cattle.
|
84 |
-
We recommend if performance is not as good as you need it to be, first try video adaptation (see Ye et al. 2023),
|
85 |
-
or fine-tune these weights with your own labeling.
|
86 |
-
|
87 |
|
88 |
## Ethical Considerations
|
89 |
|
@@ -96,8 +94,12 @@ characteristics not well-represented in the training data.
|
|
96 |
|
97 |
• Please note that each dataest was labeled by separate labs & separate individuals, therefore while we map names to a
|
98 |
unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary
|
99 |
-
Note on annotator bias).
|
100 |
-
|
|
|
|
|
|
|
|
|
101 |
good as you need it to be, first try video adaptation (see Ye et al. 2023), or fine-tune these weights with your own
|
102 |
labeling.
|
103 |
|
|
|
32 |
## Intended Use
|
33 |
• Intended to be used for pose estimation of quadruped images taken from side-view. The model serves a better starting
|
34 |
point than ImageNet weights in downstream datasets such as AP-10K.
|
35 |
+
|
36 |
• Intended for academic and research professionals working in fields related to animal behavior, such as neuroscience
|
37 |
and ecology.
|
38 |
+
|
39 |
• Not suitable as a zeros-shot model for applications that require high keypiont precision, but can be fine-tuned with
|
40 |
minimal data to reach human-level accuracy. Also not suitable for videos that look dramatically different from those
|
41 |
we show in the paper.
|
42 |
+
|
43 |
+
## Factors
|
44 |
+
|
45 |
• Based on the known robustness issues of neural networks, the relevant factors include the lighting, contrast and
|
46 |
resolution of the video frames. The present of objects might also cause false detections and erroneous keypoints.
|
47 |
When two or more animals are extremely close, it could cause the top-down detectors to only detect only one animal,
|
48 |
+
if used without further fine-tuning or with a method such as BUCTD (Zhou et al. 2023 ICCV).
|
49 |
|
50 |
## Metrics
|
51 |
• Mean Average Precision (mAP)
|
|
|
82 |
<img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690988780004-AG00N6OU1R21MZ0AU9RE/modelcard-SAQ.png?format=1500w" width="95%">
|
83 |
</p>
|
84 |
|
|
|
|
|
|
|
|
|
|
|
|
|
85 |
|
86 |
## Ethical Considerations
|
87 |
|
|
|
94 |
|
95 |
• Please note that each dataest was labeled by separate labs & separate individuals, therefore while we map names to a
|
96 |
unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary
|
97 |
+
Note on annotator bias).
|
98 |
+
|
99 |
+
• Note the dataset is highly diverse across species, but collectively has more
|
100 |
+
representation of domesticated animals like dogs, cats, horses, and cattle.
|
101 |
+
|
102 |
+
• We recommend if performance is not as
|
103 |
good as you need it to be, first try video adaptation (see Ye et al. 2023), or fine-tune these weights with your own
|
104 |
labeling.
|
105 |
|