36 projects for "slow motion video" with 1 filter applied:

  • Earn up to 15% annual interest with Nexo. Icon
    Earn up to 15% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • Earn up to 15% annual interest with Nexo. Icon
    Earn up to 15% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    SlowFast

    SlowFast

    Video understanding codebase from FAIR for reproducing video models

    SlowFast is a video understanding framework that captures both spatial semantics and temporal dynamics efficiently by processing video frames at two different temporal resolutions. The slow pathway encodes semantic context by sampling frames sparsely, while the fast pathway captures motion and fine temporal cues by operating on densely sampled frames with fewer channels.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Wan Move

    Wan Move

    Motion-controllable Video Generation via Latent Trajectory Guidance

    Wan Move is an open-source research codebase for motion-controllable video generation that focuses on enabling fine-grained control of motion within generative video models. It is designed to guide the temporal evolution of visual content by leveraging latent trajectory guidance, allowing users to manipulate how objects move over time without modifying the underlying generative architecture.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    video2robot

    video2robot

    End-to-end pipeline converting generative videos

    video2robot is an end-to-end open-source pipeline that converts generative video or prompt-driven motion content into executable humanoid robot motion sequences, enabling researchers and developers to go from high-level action descriptions or videos to robot-ready motion data. The pipeline supports both prompt-to-video generation using models like Veo/Sora and video upload processing, followed by human pose extraction through a 3D pose model and retargeting of that motion to robot joints using a general motion retargeting system. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    LTX-2.3

    LTX-2.3

    Official Python inference and LoRA trainer package

    ...This unified approach allows creators to generate complete multimedia sequences where motion, timing, and sound are aligned automatically. LTX-2 is designed for both research and production workflows and can generate high-resolution video clips with precise control over structure, motion, and camera behavior.
    Downloads: 167 This Week
    Last Update:
    See Project
  • Network Performance Monitoring | Statseeker Icon
    Network Performance Monitoring | Statseeker

    Statseeker is a powerful network performance monitoring solution for businesses

    Using just a single server or virtual machine, Statseeker can be up and running within minutes, and discovering your entire network in less than an hour, without any significant effect on your bandwidth availability.
    Learn More
  • 5
    SteadyDancer

    SteadyDancer

    Harmonized and Coherent Human Image Animation

    ...By differentiating between intentional rhythmic motion and unintentional instability, SteadyDancer applies adaptive filtering that enhances video quality without flattening the core movement dynamics. The system can be used both in preprocessing pipelines for content creators and in live feedback loops for performers, giving dancers and videographers a tool to refine their visual outputs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    The FreeMoCap Project

    The FreeMoCap Project

    Free Motion Capture for Everyone

    FreeMoCap is an open-source markerless motion capture system that enables users to record human movement using ordinary cameras and convert the footage into usable 3D motion data. The project’s goal is to democratize motion capture by removing the need for expensive suits or proprietary studio hardware, instead relying on computer vision and pose estimation pipelines. It processes synchronized video feeds to reconstruct skeletal motion, which can then be exported for animation, biomechanics research, or creative projects. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 7
    WanGP

    WanGP

    AI video generator optimized for low VRAM and older GPUs use

    ...Wan2GP provides a full web-based interface that simplifies interaction with complex generative pipelines, making it easier to configure prompts, models, and rendering settings. It also integrates a wide range of utilities such as prompt enhancement, mask editing, motion design, and extraction tools for pose, depth, and flow data to support advanced video workflows.
    Downloads: 33 This Week
    Last Update:
    See Project
  • 8
    Remotion

    Remotion

    Make videos programmatically with React

    Remotion is a cutting-edge library that lets developers create real videos programmatically using React components, transforming familiar UI paradigms into a flexible, code-driven video production workflow. Instead of traditional timeline editors, Remotion leverages HTML, CSS, and JavaScript to define video frames, animations, and transitions, which means developers can use states, props, loops, and component hierarchies to automate complex motion graphics. Because it integrates with the React ecosystem, Remotion fits naturally into modern front-end stacks and tooling, and can produce dynamic content like personalized videos, dashboards, and data-driven animations with the same code used to build interactive web apps. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 9
    Peinture

    Peinture

    A general-purpose AI image generation framework that supports HF

    Peinture is a sleek, dark-themed web application that brings AI-powered image generation to artists, designers, and casual creators through a modern interface built with React, TypeScript, and Tailwind CSS. Instead of tying users to a single service, Peinture integrates multiple backend providers, including Hugging Face, Gitee AI, Model Scope, and others, so you can switch between models or extend support to custom endpoints — giving you flexibility over where and how your images are...
    Downloads: 0 This Week
    Last Update:
    See Project
  • PeerGFS PEER Software - File Sharing and Collaboration Icon
    PeerGFS PEER Software - File Sharing and Collaboration

    One Solution to Simplify File Management and Orchestration Across Edge, Data Center, and Cloud Storage

    PeerGFS is a software-only solution developed to solve file management/file replication challenges in multi-site, multi-platform, and hybrid multi-cloud environments.
    Learn More
  • 10
    The Arcade Library

    The Arcade Library

    Easy to use Python library for creating 2D arcade games

    Arcade is an easy-to-use Python library for creating 2D video games. It provides a modern and straightforward API, enabling developers to craft engaging games and graphical applications efficiently. Arcade supports rendering shapes, handling user input, and managing game physics, making it suitable for both beginners and experienced developers.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 11
    HunyuanVideo-Avatar

    HunyuanVideo-Avatar

    Tencent Hunyuan Multimodal diffusion transformer (MM-DiT) model

    ...Emotion control by extracting emotion reference images and transferring emotional style into video sequences.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    LiveAvatar

    LiveAvatar

    Streaming Real-time Audio-Driven Avatar Generation

    LiveAvatar is an open-source research and implementation project that provides a unified framework for real-time, streaming, interactive avatar video generation driven by audio and other control signals. It implements techniques from state-of-the-art diffusion-based avatar modeling to support infinite-length continuous video generation with low latency, enabling interactive AI avatars that maintain continuity and realism over extended sessions. The project co-designs algorithms and system...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Frigate

    Frigate

    NVR with realtime local object detection for IP cameras

    Frigate - NVR With Realtime Object Detection for IP Cameras A complete and local NVR designed for Home Assistant with AI object detection. Uses OpenCV and Tensorflow to perform realtime object detection locally for IP cameras. Use of a Google Coral Accelerator is optional, but highly recommended. The Coral will outperform even the best CPUs and can process 100+ FPS with very little overhead.
    Downloads: 35 This Week
    Last Update:
    See Project
  • 14
    vJEPA-2

    vJEPA-2

    PyTorch code and models for VJEPA2 self-supervised learning from video

    VJEPA2 is a next-generation self-supervised learning framework for video that extends the “predict in representation space” idea from i-JEPA to the temporal domain. Instead of reconstructing pixels, it predicts the missing high-level embeddings of masked space-time regions using a context encoder and a slowly updated target encoder. This objective encourages the model to learn semantics, motion, and long-range structure without the shortcuts that pixel-level losses can invite. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    BoxMOT

    BoxMOT

    Pluggable SOTA multi-object tracking modules for segmentation

    BoxMOT is an open-source framework designed to provide modular implementations of state-of-the-art multi-object tracking algorithms for computer vision applications. The project focuses on the tracking-by-detection paradigm, where objects detected by vision models are continuously tracked across frames in a video sequence. It provides a pluggable architecture that allows developers to combine different object detectors with multiple tracking algorithms without modifying the core codebase....
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    VideoCrafter2

    VideoCrafter2

    Overcoming Data Limitations for High-Quality Video Diffusion Models

    VideoCrafter is an open-source video generation and editing toolbox designed to create high-quality video content. It features models for both text-to-video and image-to-video generation. The system is optimized for generating videos from textual descriptions or still images, leveraging advanced diffusion models. VideoCrafter2, an upgraded version, improves on its predecessor by enhancing motion dynamics and concept combinations, especially in low-data scenarios. ...
    Downloads: 16 This Week
    Last Update:
    See Project
  • 17
    MLT Multimedia Framework
    A multimedia authoring and processing framework and a video playout server for television broadcasting.
    Downloads: 30 This Week
    Last Update:
    See Project
  • 18
    jfRecordCamera

    jfRecordCamera

    Camera recorder

    Records your camera/mic to a video file. Includes preview while recording. Time lapse and stop motion modes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    CoTracker

    CoTracker

    CoTracker is a model for tracking any point (pixel) on a video

    ...The repository ships with inference scripts, pretrained weights, and simple interfaces to seed points, run tracking, and export trajectories for downstream tasks. Typical uses include correspondence building, motion analysis, dynamic SLAM priors, video editing masks, and evaluation of geometric consistency in real scenes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Gait Tracking With x-IMU

    Gait Tracking With x-IMU

    Foot tracking algorithm demonstrated in Seb Madgwick's 3D Tracking

    This repository provides the source code for a foot-tracking algorithm using an inertial measurement unit (IMU) attached to the foot, as demonstrated in Seb Madgwick’s “3D Tracking with IMU” video (March 2011). The method performs dead reckoning to integrate foot motion and corrects for drift at each foot strike (when the foot touches the ground) to reset error accumulation. The algorithm was part of gait analysis research, estimating foot trajectories in 3D over time. The repo is marked “no longer maintained,” and has been superseded by newer versions of Gait-Tracking. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    FrankMocap

    FrankMocap

    A Strong and Easy-to-use Single View 3D Hand+Body Pose Estimator

    ...Outputs include textured meshes, joint locations, and model parameters that can be exported to common DCC tools and game engines. The codebase offers pretrained models, clear inference scripts, and utilities to visualize results, making single-camera motion capture approachable on commodity hardware. Researchers and creators use it for motion studies, AR/VR prototyping, character animation, and human-in-the-loop editing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Motionity

    Motionity

    The web-based motion graphics editor for everyone

    Motionity is an open-source, browser-based motion-graphics editor designed to bring animation and video editing power to the web, blending the intuitiveness of drag-and-drop tools with advanced capabilities akin to desktop motion editors. It allows users to assemble video or image assets, layer them, and apply animation via keyframes — with customizable easing — giving fine-grained control over motion timing.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    TimeSformer

    TimeSformer

    The official pytorch implementation of our paper

    TimeSformer is a vision transformer architecture for video that extends the standard attention mechanism into spatiotemporal attention. The model alternates attention along spatial and temporal dimensions (or designs variants like divided attention) so that it can capture both appearance and motion cues in video. Because the attention is global across frames, TimeSformer can reason about dependencies across long time spans, not just local neighborhoods.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    Nerfies

    Nerfies

    This is the code for Deformable Neural Radiance Fields

    ...A set of utilities manages dataset preparation, pose estimation, and checkpoints so researchers can reproduce results on their own footage. The work sits at the intersection of graphics and vision, showing how learned volumetric rendering can handle human motion without dense markers or studio rigs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    DensePose

    DensePose

    A real-time approach for mapping all human pixels of 2D RGB images

    DensePose is a computer vision system that maps all human pixels in an RGB image to the 3D surface of a human body model. It extends human pose estimation from predicting joint keypoints to providing dense correspondences between 2D images and a canonical 3D mesh (such as the SMPL model). This enables detailed understanding of human shape, motion, and surface appearance directly from images or videos. The repository includes the DensePose network architecture, training code, pretrained...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB