mwmathis commited on
Commit
57e3f36
1 Parent(s): 03264fc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -10
README.md CHANGED
@@ -32,16 +32,20 @@ download_huggingface_model("superanimal_quadruped", model_dir)
32
  ## Intended Use
33
  • Intended to be used for pose estimation of quadruped images taken from side-view. The model serves a better starting
34
  point than ImageNet weights in downstream datasets such as AP-10K.
 
35
  • Intended for academic and research professionals working in fields related to animal behavior, such as neuroscience
36
  and ecology.
 
37
  • Not suitable as a zeros-shot model for applications that require high keypiont precision, but can be fine-tuned with
38
  minimal data to reach human-level accuracy. Also not suitable for videos that look dramatically different from those
39
  we show in the paper.
40
- Factors
 
 
41
  • Based on the known robustness issues of neural networks, the relevant factors include the lighting, contrast and
42
  resolution of the video frames. The present of objects might also cause false detections and erroneous keypoints.
43
  When two or more animals are extremely close, it could cause the top-down detectors to only detect only one animal,
44
- if used without further fine-tuning or with a method such as BUCTD (36).
45
 
46
  ## Metrics
47
  • Mean Average Precision (mAP)
@@ -78,12 +82,6 @@ Here is an image with the keypoint guide:
78
  <img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690988780004-AG00N6OU1R21MZ0AU9RE/modelcard-SAQ.png?format=1500w" width="95%">
79
  </p>
80
 
81
- Please note that each dataset was labeled by separate labs \& separate individuals, therefore while we map names
82
- to a unified pose vocabulary (found here: https://github.com/AdaptiveMotorControlLab/modelzoo-figures), there will be annotator bias in keypoint placement (See the Supplementary Note on annotator bias).
83
- You will also note the dataset is highly diverse across species, but collectively has more representation of domesticated animals like dogs, cats, horses, and cattle.
84
- We recommend if performance is not as good as you need it to be, first try video adaptation (see Ye et al. 2023),
85
- or fine-tune these weights with your own labeling.
86
-
87
 
88
  ## Ethical Considerations
89
 
@@ -96,8 +94,12 @@ characteristics not well-represented in the training data.
96
 
97
  • Please note that each dataest was labeled by separate labs & separate individuals, therefore while we map names to a
98
  unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary
99
- Note on annotator bias). You will also note the dataset is highly diverse across species, but collectively has more
100
- representation of domesticated animals like dogs, cats, horses, and cattle. We recommend if performance is not as
 
 
 
 
101
  good as you need it to be, first try video adaptation (see Ye et al. 2023), or fine-tune these weights with your own
102
  labeling.
103
 
 
32
  ## Intended Use
33
  • Intended to be used for pose estimation of quadruped images taken from side-view. The model serves a better starting
34
  point than ImageNet weights in downstream datasets such as AP-10K.
35
+
36
  • Intended for academic and research professionals working in fields related to animal behavior, such as neuroscience
37
  and ecology.
38
+
39
  • Not suitable as a zeros-shot model for applications that require high keypiont precision, but can be fine-tuned with
40
  minimal data to reach human-level accuracy. Also not suitable for videos that look dramatically different from those
41
  we show in the paper.
42
+
43
+ ## Factors
44
+
45
  • Based on the known robustness issues of neural networks, the relevant factors include the lighting, contrast and
46
  resolution of the video frames. The present of objects might also cause false detections and erroneous keypoints.
47
  When two or more animals are extremely close, it could cause the top-down detectors to only detect only one animal,
48
+ if used without further fine-tuning or with a method such as BUCTD (Zhou et al. 2023 ICCV).
49
 
50
  ## Metrics
51
  • Mean Average Precision (mAP)
 
82
  <img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690988780004-AG00N6OU1R21MZ0AU9RE/modelcard-SAQ.png?format=1500w" width="95%">
83
  </p>
84
 
 
 
 
 
 
 
85
 
86
  ## Ethical Considerations
87
 
 
94
 
95
  • Please note that each dataest was labeled by separate labs & separate individuals, therefore while we map names to a
96
  unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary
97
+ Note on annotator bias).
98
+
99
+ • Note the dataset is highly diverse across species, but collectively has more
100
+ representation of domesticated animals like dogs, cats, horses, and cattle.
101
+
102
+ • We recommend if performance is not as
103
  good as you need it to be, first try video adaptation (see Ye et al. 2023), or fine-tune these weights with your own
104
  labeling.
105