The napari hub is transitioning to a community-run implementation due to launch in June 2025.
Since October 1, 2024, this version is no longer actively maintained and will not be updated. New plugins and plugin updates will continue to be listed.

movement

movement

Analysis of body movement

    Python Version PyPI Version Conda Forge Version License CI codecov Binder Code style: Ruff pre-commit project chat DOI

    A Python toolbox for analysing animal body movements across space and time.

    Quick install

    Create and activate a conda environment with movement installed (including the GUI):

    conda create -n movement-env -c conda-forge movement napari pyqt
    conda activate movement-env

    [!Note] Read the documentation for more information, including full installation instructions and examples.

    Overview

    Deep learning methods for motion tracking have revolutionised a range of scientific disciplines, from neuroscience and biomechanics, to conservation and ethology. Tools such as DeepLabCut and SLEAP now allow researchers to track animal movements in videos with remarkable accuracy, without requiring physical markers. However, there is still a need for standardised, easy-to-use methods to process the tracks generated by these tools.

    movement aims to provide a consistent, modular interface for analysing motion tracks, enabling steps such as data cleaning, visualisation, and motion quantification. We aim to support all popular animal tracking frameworks and file formats.

    Find out more on our mission and scope statement and our roadmap.

    [!Warning] 🏗️ The package is currently in early development and the interface is subject to change. Feel free to play around and provide feedback.

    [!Tip] If you prefer analysing your data in R, we recommend checking out the animovement toolbox, which is similar in scope. We are working together with its developer to gradually converge on common data standards and workflows.

    Join the movement

    Contributions to movement are absolutely encouraged, whether to fix a bug, develop a new feature, or improve the documentation. To help you get started, we have prepared a detailed contributing guide.

    Citation

    If you use movement in your work, please cite the following Zenodo DOI:

    Nikoloz Sirmpilatze, Chang Huan Lo, Sofía Miñano, Brandon D. Peri, Dhruv Sharma, Laura Porta, Iván Varela & Adam L. Tyson (2024). neuroinformatics-unit/movement. Zenodo. https://zenodo.org/doi/10.5281/zenodo.12755724

    License

    ⚖️ BSD 3-Clause

    Package template

    This package layout and configuration (including pre-commit hooks and GitHub actions) have been copied from the python-cookiecutter template.

    Version:

    • 0.2.0

    Last updated:

    • 10 March 2025

    First released:

    • 02 April 2020

    License:

    Supported data:

    • Information not submitted

    Plugin type:

    GitHub activity:

    • Stars: 138
    • Forks: 36
    • Issues + PRs: 125

    Python versions supported:

    Operating system:

    Requirements:

    • numpy
    • pandas
    • h5py
    • attrs
    • pooch
    • tqdm
    • shapely
    • sleap-io
    • xarray[accel,viz]
    • PyYAML
    • napari-video
    • pyvideoreader>=0.5.3
    • qt-niu