Edit model card

MODEL CARD:

Model Details

• SuperAnimal-TopViewMouse model developed by the M.W.Mathis Lab in 2023, trained to predict mouse topline pose from images. Please see Shaokai Ye et al. 2023 for details.

• The there are three models:

  • pose_model.pth is an HRNet-w32 compatable for DLC3.0+ Pytorch code, trained on our TopViewMouse-5K dataset.
  • detector.pt is a Faster R-CNN that can be used as a detector for top-down detection.
  • DLC_ma_supertopview5k_resnet_50_iteration-0_shuffle-1.tar.gz is a DLCRNet trained on our TopViewMouse-5K dataset.

• Full training details can be found in Ye et al. 2023. You can use this model simply with our light-weight loading package called DLCLibrary. Here is an example useage:

from pathlib import Path
from dlclibrary import download_huggingface_model

# Creates a folder and downloads the model to it
model_dir = Path("./superanimal_topviewmouse_model")
model_dir.mkdir()
download_huggingface_model("superanimal_topviewmouse_model", model_dir)

Intended Use

• Intended to be used for pose tracking of lab mice videos filmed from an overhead view. The models can be used as a plug-and- play solution if extremely high precision is not required (we benchmark the zero-shot performance in the paper). Otherwise, it is recommended to also be used as the weights for transfer learning and fine-tuning.

• Intended for academic and research professionals working in fields related to animal behavior, neuroscience, biomechanics, and ecology.

• Not suitable for other species and other camera views. Also not suitable for videos that look dramatically different from those we show in the paper.

Factors

• Based on the known robustness issues of neural networks, the relevant factors include the lighting, contrast and resolution of the video frames. The present of objects might also cause false detections of the mice and keypoints. When two or more animals are extremely close, it could cause the top-down detectors to only detect only one animal, if used without further fine-tuning.

Metrics

• Mean Average Precision (mAP)

• Root Mean Square Error (RMSE)

Evaluation Data

• The test split of TopViewMouse-5K and in the paper on two benchmarks, DLC Openfield and TriMouse

Training Data

It consists of being trained together on the following datasets:

  • 3CSI, BM, EPM, LDB, OFT See full details at (1) and in (2).

  • BlackMice See full details at (3).

  • WhiteMice Courtesy of Prof. Sam Golden and Nastacia Goodwin. See details in SIMBA (4). TriMouse See full details at (5).

  • DLC-Openfield See full details at (6).

  • Kiehn-Lab-Openfield, Swimming, and treadmill Courtesy of Prof. Ole Kiehn, Dr. Jared Cregg, and Prof. Carmelo Bellardita; see details at (7).

  • MausHaus We collected video data from five single-housed C57BL/6J male and female mice in an extended home cage, carried out in the laboratory of Mackenzie Mathis at Harvard University and also EPFL (temperature of housing was 20-25C, humidity 20-50%). Data were recorded at 30Hz with 640 × 480 pixels resolution acquired with White Matter, LLC eV cameras. Annotators localized 26 keypoints across 322 frames sampled from within DeepLabCut using the k-means clustering approach (8). All experimental procedures for mice were in accordance with the National Institutes of Health Guide for the Care and Use of Laboratory Animals and approved by the Harvard Institutional Animal Care and Use Committee (IACUC) (n=1 mouse), and by the Veterinary Office of the Canton of Geneva (Switzerland; license GE01) (n=4 mice).

Here is an image with examples from the datasets, the distribution of images per dataset, and the keypoint guide.

Ethical Considerations

• Data was collected with IUCAC or other governmental approval. Each individual dataset used in training reports the ethics approval they obtained.

Caveats and Recommendations

• The model may have reduced accuracy in scenarios with extremely varied lighting conditions or atypical mouse characteristics not well-represented in the training data. For example, this dataset only has one set of white mice, therefore it may not generalize well to diverse settings of white lab mice.

• Please note that each training dataset was labeled by separate labs and different individuals, therefore while we map names to a unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary Note on annotator bias).

• Note the dataset is primarily using C56Blk6/J mice and only some CD1 examples.

• We recommend if performance is not as good as you need it to be, first try video adaptation (see Ye et al. 2023), or fine-tune these weights with your own labeling.

License

Modified MIT.

Copyright 2023 by Mackenzie Mathis, Shaokai Ye, and contributors.

Permission is hereby granted to you (hereafter "LICENSEE") a fully-paid, non-exclusive, and non-transferable license for academic, non-commercial purposes only (hereafter “LICENSE”) to use the "MODEL" weights (hereafter "MODEL"), subject to the following conditions:

The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software:

This software may not be used to harm any animal deliberately.

LICENSEE acknowledges that the MODEL is a research tool. THE MODEL IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE MODEL OR THE USE OR OTHER DEALINGS IN THE MODEL.

If this license is not appropriate for your application, please contact Prof. Mackenzie W. Mathis (mackenzie@post.harvard.edu) and/or the TTO office at EPFL (tto@epfl.ch) for a commercial use license.

Please cite Ye et al if you use this model in your work https://arxiv.org/abs/2203.07436v2.

References

  1. Oliver Sturman, Lukas von Ziegler, Christa Schläppi, Furkan Akyol, Mattia Privitera, Daria Slominski, Christina Grimm, Laetitia Thieren, Valerio Zerbi, Benjamin Grewe, et al. Deep learning-based behavioral analysis reaches human accuracy and is capable of outperforming commercial solutions. Neuropsychopharmacology, 45(11):1942–1952, 2020.
  2. Lukas von Ziegler, Oliver Sturman, and Johannes Bohacek. Videos for deeplabcut, noldus ethovision X14 and TSE multi conditioning systems comparisons. https://doi.org/10.5281/zenodo.3608658. Zenodo, January 2020.
  3. Isaac Chang. Trained DeepLabCut model for tracking mouse in open field arena with topdown view. https://doi.org/10.5281/zenodo.3955216. Zenodo, July 2020.
  4. Simon RO Nilsson, Nastacia L. Goodwin, Jia Jie Choong, Sophia Hwang, Hayden R Wright, Zane C Norville, Xiaoyu Tong, Dayu Lin, Bran- don S. Bentzley, Neir Eshel, Ryan J McLaughlin, and Sam A. Golden. Simple behavioral analysis (simba) – an open source toolkit for computer classification of complex social behaviors in experimental animals. bioRxiv, 2020.
  5. Jessy Lauer, Mu Zhou, Shaokai Ye, William Menegas, Steffen Schneider, Tanmay Nath, Mohammed Mostafizur Rahman, Valentina Di Santo, Daniel Soberanes, Guoping Feng, Venkatesh N. Murthy, George Lauder, Catherine Dulac, Mackenzie W. Mathis, and Alexander Mathis. Multi- animal pose estimation, identification and tracking with deeplabcut. Nature Methods, 19:496 – 504, 2022.
  6. Alexander Mathis, Pranav Mamidanna, Kevin M Cury, Taiga Abe, Venkatesh N Murthy, Mackenzie Weygandt Mathis, and Matthias Bethge. Deeplab- cut: markerless pose estimation of user-defined body parts with deep learning. Nature neuroscience, 21:1281–1289, 2018.
  7. Jared M. Cregg, Roberto Leiras, Alexia Montalant, Paulina Wanken, Ian R. Wickersham, and Ole Kiehn. Brainstem neurons that command mammalian locomotor asymmetries. Nature neuroscience, 23:730 – 740, 2020
Downloads last month
0
Unable to determine this model's library. Check the docs .