Showing 19 open source projects for "image sequence"

View related business solutions
  • Earn up to 15% annual interest with Nexo. Icon
    Earn up to 15% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • Earn up to 15% annual interest with Nexo. Icon
    Earn up to 15% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    HunyuanVideo-I2V

    HunyuanVideo-I2V

    A Customizable Image-to-Video Model based on HunyuanVideo

    HunyuanVideo-I2V is a customizable image-to-video generation framework from Tencent Hunyuan, built on their HunyuanVideo foundation. It extends video generation so that given a static reference image plus an optional prompt, it generates a video sequence that preserves the reference image’s identity (especially in the first frame) and allows stylized effects via LoRA adapters.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    PersonaLive

    PersonaLive

    Expressive Portrait Image Animation for Live Streaming

    PersonaLive is an open-source diffusion-based portrait animation framework focused on generating expressive, long-duration animated sequences in real time, primarily for live streaming or interactive applications. It leverages deep generative models that condition on a static reference image and a driving input (such as motion or expression cues) to produce a seamless animated portrait sequence that can run indefinitely without segmentation artifacts. The framework prioritizes low-latency and streamable output, making it suitable for real-time creative workflows, broadcast overlays, or interactive avatars on consumer-grade GPUs. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    Lingvo

    Lingvo

    Framework for building neural networks

    Lingvo is a TensorFlow based framework focused on building and training sequence models, especially for language and speech tasks. It was originally developed for internal research and later open sourced to support reproducible experiments and shared model implementations. The framework provides a structured way to define models, input pipelines, and training configurations using a common interface for layers, which encourages reuse across different tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    ktrain

    ktrain

    ktrain is a Python library that makes deep learning AI more accessible

    ktrain is a Python library that makes deep learning and AI more accessible and easier to apply. ktrain is a lightweight wrapper for the deep learning library TensorFlow Keras (and other libraries) to help build, train, and deploy neural networks and other machine learning models. Inspired by ML framework extensions like fastai and ludwig, ktrain is designed to make deep learning and AI more accessible and easier to apply for both newcomers and experienced practitioners. With only a few lines...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Iris Powered By Generali - Iris puts your customer in control of their identity. Icon
    Iris Powered By Generali - Iris puts your customer in control of their identity.

    Increase customer and employee retention by offering Onwatch identity protection today.

    Iris Identity Protection API sends identity monitoring and alerts data into your existing digital environment – an ideal solution for businesses that are looking to offer their customers identity protection services without having to build a new product or app from scratch.
    Learn More
  • 5
    HunyuanVideo-I2V

    HunyuanVideo-I2V

    A Customizable Image-to-Video Model based on HunyuanVideo

    HunyuanVideo-I2V is a customizable image-to-video generation framework developed by Tencent, extending the capabilities of HunyuanVideo. It allows for high-quality video creation from still images, using PyTorch and providing pre-trained model weights, inference code, and customizable training options. The system includes a LoRA training code for adding special effects and enhancing video realism, aiming to offer versatile and scalable solutions for generating videos from static image inputs.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 6
    PulseClickV1

    PulseClickV1

    Autoclicker

    ...Featuring a stunning animated neon particle background, glassmorphism effects, glowing buttons, and smooth cyberpunk aesthetics. ### Key Features - **10 Powerful Clicking Modes**: Classic, Color Click, Picture/Image Click, Random Burst, Swipe/Drag, Hold & Release, Jitter Click, Pixel Scan, Timed Sequence, and Smart Region. - **Advanced Screen Area Selection**: Hold Ctrl + drag anywhere on your screen to select a custom click region. - **Multi-Click & Burst Support**: Choose 1, 5, 10, 50 clicks or continuous hold mode. - **Hotkey Control**: Press F6 to start/stop clicking instantly...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    IPyPlot

    IPyPlot

    Fast and efficient plotting of images inside Python Notebooks

    IPyPlot is a small python package offering fast and efficient plotting of images inside Python Notebooks. It's using IPython with HTML for faster, richer and more interactive way of displaying big numbers of images.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    DALL-E in Pytorch

    DALL-E in Pytorch

    Implementation / replication of DALL-E, OpenAI's Text to Image

    ...Currently only the VAE with a codebook size of 1024 is offered, with the hope that it may train a little faster than OpenAI's, which has a size of 8192. In contrast to OpenAI's VAE, it also has an extra layer of downsampling, so the image sequence length is 256 instead of 1024 (this will lead to a 16 reduction in training costs, when you do the math).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Amiga Memories

    Amiga Memories

    A walk along memory lane

    ...The purpose of the script is to define the spoken and written content. The spoken text will be read by a voice synthesizer (Text To Speech or TTS), the written text is simply drawn on the image as subtitles. Here, in addition to the spoken & written narration, the script controls the camera movements as well as the LED activity of the computer. Amiga Memories' video images are computed by the GameStart 3D engine (pre-HARFANG 3D). Although the 3D assets are designed to be played back in real-time with a variable framerate, the engine is capable of breaking down the video sequence into the 30th or 60th of a second, as TGA files.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • 10
    Minimal text diffusion

    Minimal text diffusion

    A minimal implementation of diffusion models for text generation

    A minimal implementation of diffusion models of text: learns a diffusion model of a given text corpus, allowing to generate text samples from the learned model. The main idea was to retain just enough code to allow training a simple diffusion model and generating samples, remove image-related terms, and make it easier to use. To train a model, run scripts/train.sh. By default, this will train a model on the simple corpus. However, you can change this to any text file using the --train_data argument. Note that you may have to increase the sequence length (--seq_len) if your corpus is longer than the simple corpus. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    RQ-Transformer

    RQ-Transformer

    Implementation of RQ Transformer, autoregressive image generation

    ...I also think there is something deeper going on, and have generalized this to any number of dimensions. You can use it by importing the HierarchicalCausalTransformer. For autoregressive (AR) modeling of high-resolution images, vector quantization (VQ) represents an image as a sequence of discrete codes. A short sequence length is important for an AR model to reduce its computational costs to consider long-range interactions of codes. However, we postulate that previous VQ cannot shorten the code sequence and generate high-fidelity images together in terms of the rate-distortion trade-off.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Deep Daze

    Deep Daze

    Simple command line tool for text to image generation

    Simple command-line tool for text to image generation using OpenAI's CLIP and Siren (Implicit neural representation network). In true deep learning fashion, more layers will yield better results. Default is at 16, but can be increased to 32 depending on your resources. Technique first devised and shared by Mario Klingemann, it allows you to prime the generator network with a starting image, before being steered towards the text. Simply specify the path to the image you wish to use, and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Deep Exemplar-based Video Colorization

    Deep Exemplar-based Video Colorization

    The source code of CVPR 2019 paper "Deep Exemplar-based Colorization"

    ...To address this issue, we introduce a recurrent framework that unifies the semantic correspondence and color propagation steps. Both steps allow a provided reference image to guide the colorization of every frame, thus reducing accumulated propagation errors. Video frames are colorized in sequence based on the colorization history, and its coherency is further enforced by the temporal consistency loss. All of these components, learned end-to-end, help produce realistic videos with good temporal stability. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Sparse Attention

    Sparse Attention

    "Generating Long Sequences with Sparse Transformers" examples

    ...It highlights both fixed and learnable sparsity patterns that trade off computational cost and model expressiveness. By enabling tractable training on longer contexts, the project opened the door to applications in large-scale text and image generation. Though archived, it remains a key reference for efficient transformer research, influencing many later architectures that aim to extend sequence length while reducing compute.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    DCVGAN

    DCVGAN

    DCVGAN: Depth Conditional Video Generation, ICIP 2019.

    This paper proposes a new GAN architecture for video generation with depth videos and color videos. The proposed model explicitly uses the information of depth in a video sequence as additional information for a GAN-based video generation scheme to make the model understands scene dynamics more accurately. The model uses pairs of color video and depth video for training and generates a video using the two steps. Generate the depth video to model the scene dynamics based on the geometrical information. To add appropriate color to the geometrical information of the scene, the domain translation from depth to color is performed for each image. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    SPADE

    SPADE

    A toolkit for developing and deploying protein structure algorithms.

    The Structural Proteomics Application Development Environment is a Python tool kit for developing and deploying bioinformatics applications. Handles graphics, analysis, and modeling of protein sequence and structure. Source and Win installers available. SPADE source code can be cloned from http://www.github.com/deaconjs/SPADE.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Tool for annotating (bike race) videos with data from GPS devices to display time, distance, speed, power, HR, CAD, elevation, laps. Outputs a series of SVG files importable as image sequence into the OpenShot Video editor as overplayed video track.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    bni-tools
    --UPDATE-- New Version is now integrated in the official PyMOL plugin repository! --UPDATE-- The BNI (Beyond Normal Interaction)- Tools is a plug in for the PyMOL molecular visualization system which adds additional functionalities and presets to the PyMOL GUI and also adds useful extended commands.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Comic reader is a comic book reader written in Python. Its meant to run on linux and browse through image archives in ZIP or RAR format. The program extracts the pages in sequence and displays them using an internal viewer or an external program.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB