The CameraTrapDetector project is a set of customized object detection deep learning models that identify, classify, and count animals in camera trap images. The model can be run on personal computer as part of an existing workflow, saving time and preserving data privacy.
Our mission is to provide an accurate, easy-to-use, free computer vision tool to process large camera trap datasets. Users with no coding experience can automate the time-intensive task of classifying images, and optimize their time spent thinking analytically. Our tool enables researchers and land managers to perform analyses and make decisions faster and with more comprehensive information.
🐷 To use the packaged model in R, RStudio, or the R Shiny interface, see the CameraTrapDetectoR repository.
🐮 To use the interactive app on your desktop without any code interface, see the Desktop App repository.
🐇 To deploy the models on HPC or via command line, see the Model Training repository.
🌲 See our new package on creating your own randomly-placed camera trap array: cameratrapgridR
🐴 To peruse the helper and template scripts related to our project, see the Resources repository.
🐖 To collaborate with us by sharing images or offering feedback, send us a message: [email protected]
CameraTrapDetector hosts custom-trained deep object detection models at the taxonomic class, family, and species levels. The latest published version of each model is Version 2
Track our progress as we work to advance CameraTrapDetector's mission of publishing new, more accurate models with multiple, accessible use options.
Data curation:
✔️ Detection-level image features of contrast, complexity, self-similarity, and symmetry to our annotations for better representative sampling📈 Animal-level orientation, size, distance annotation for better representative sampling
📈 Incorporate new images into training database
🔲 Generate images of rare/invasive species in various settings and orientations using SMOTE techniques
📈 Review existing database to flag poor quality training samples
Model Training:
✔️ Pretrain Faster-RCNN weights on OOS camera trap images✔️ Pretrain YOLO weights on OOS camera trap images
📈 transfer learning with pre-existing camera trap model weights
📈 Faster-RCNN model comparison with differing CNN backbones
🔲 Prediction on video files
R Package Deployment:
✔️ Prediction verification toolkit🔲 R Shiny makeover
🔲 Submit package to CRAN for greater visibility
Desktop Deployment:
🔲 Revamp approach for more nimble appCommand Line / HPC Deployment:
✔️ Standardize output, including prediction plots, to R package output🔲 cross-platform generalizability via containerization
🔲 Integrate parallel processing, multi-GPU deployment
Evaluation / Verification Workflow:
✔️ Output compatibility with [Camelot](https://camelotproject.org/)📈 Model mis-classified images to determine feature association with prediction accuracy
Each model type comes with its own DOI. If you use CameraTrapDetectoR in your work, please cite the model type(s) you deploy:
Species Model:
@article{Burns2023, author = "Amira Burns and Ryan Miller and Hailey Wilmer and Michael Tabak and Daniel Falbel and Tess Hamzeh and Ryan K. Brook and John A. Goolsby and Lisa D. Zoromski and Raoul Boughton and Nathan Snow and Kurt VerCauteren", title = "{CameraTrapDetectoR Species Model}", year = "2023", month = "5", url = "https://agdatacommons.nal.usda.gov/articles/model/CameraTrapDetectoR_Species_Model/25234231", doi = "10.15482/USDA.ADC/1528955" }
Family Model:
@article{Burns2023, author = "Amira Burns and Michael Tabak and Daniel Falbel and Tess Hamzeh and Ryan K. Brook and John A. Goolsby and Lisa D. Zoromski and Raoul Boughton and Nathan Snow and Kurt VerCauteren and Ryan Miller and Hailey Wilmer", title = "{CameraTrapDetectoR Family Model}", year = "2023", month = "5", url = "https://agdatacommons.nal.usda.gov/articles/model/CameraTrapDetectoR_Family_Model/25234243", doi = "10.15482/USDA.ADC/1528979" }
General Model:
@article{Burns2023, author = "Amira Burns and Ryan Miller and Hailey Wilmer and Michael Tabak and Daniel Falbel and Tess Hamzeh and Ryan K. Brook and John A. Goolsby and Lisa D. Zoromski and Raoul Boughton and Nathan Snow and Kurt VerCauteren", title = "{CameraTrapDetectoR General Model}", year = "2023", month = "5", url = "https://agdatacommons.nal.usda.gov/articles/model/CameraTrapDetectoR_General_Model/25234258", doi = "10.15482/USDA.ADC/1528970" }
To cite our supporting paper:
Tabak, M. A., Falbel, D., Hamzeh, T., Brook, R. K., Goolsby, J. A., Zoromski, L. D., Boughton, R. K., Snow, N. P., VerCauteren, K. C., & Miller, R. S. (2022). CameraTrapDetectoR: Automatically detect, classify, and count animals in camera trap images using artificial intelligence (p. 2022.02.07.479461). bioRxiv. link to manuscript
Or
@article {Tabak2022.02.07.479461, author = {Tabak, Michael A and Falbel, Daniel and Hamzeh, Tess and Brook, Ryan K and Goolsby, John A and Zoromski, Lisa D and Boughton, Raoul K and Snow, Nathan P and VerCauteren, Kurt C and Miller, Ryan S}, title = {CameraTrapDetectoR: Automatically detect, classify, and count animals in camera trap images using artificial intelligence}, elocation-id = {2022.02.07.479461}, year = {2022}, doi = {10.1101/2022.02.07.479461}, publisher = {Cold Spring Harbor Laboratory},, URL = {https://www.biorxiv.org/content/10.1101/2022.02.07.479461v1}, eprint = {https://www.biorxiv.org/content/10.1101/2022.02.07.479461v1.full.pdf}, journal = {bioRxiv} }