Open Source Linux Deep Learning Frameworks - Page 3

Deep Learning Frameworks for Linux

View 20 business solutions
  • Earn up to 15% annual interest with Nexo. Icon
    Earn up to 15% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • Earn up to 15% annual interest with Nexo. Icon
    Earn up to 15% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    PaddleX

    PaddleX

    PaddlePaddle End-to-End Development Toolkit

    PaddleX is a deep learning full-process development tool based on the core framework, development kit, and tool components of Paddle. It has three characteristics opening up the whole process, integrating industrial practice, and being easy to use and integrate. Image classification and labeling is the most basic and simplest labeling task. Users only need to put pictures belonging to the same category in the same folder. When the model is trained, we need to divide the training set, the validation set and the test set. Therefore, we need to divide the above data. Using the paddlex command, the data set can be randomly divided into 70% training set, 20% validation set and 10% test set. If you use the PaddleX visualization client for model training, the data set division function is integrated in the client, and you do not need to use command division by yourself.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Ray

    Ray

    A unified framework for scalable computing

    Modern workloads like deep learning and hyperparameter tuning are compute-intensive and require distributed or parallel execution. Ray makes it effortless to parallelize single machine code — go from a single CPU to multi-core, multi-GPU or multi-node with minimal code changes. Accelerate your PyTorch and Tensorflow workload with a more resource-efficient and flexible distributed execution framework powered by Ray. Accelerate your hyperparameter search workloads with Ray Tune. Find the best model and reduce training costs by using the latest optimization algorithms. Deploy your machine learning models at scale with Ray Serve, a Python-first and framework agnostic model serving framework. Scale reinforcement learning (RL) with RLlib, a framework-agnostic RL library that ships with 30+ cutting-edge RL algorithms including A3C, DQN, and PPO. Easily build out scalable, distributed systems in Python with simple and composable primitives in Ray Core.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    Robust Tube MPC

    Robust Tube MPC

    Example implementation for robust model predictive control using tube

    robust-tube-mpc is a MATLAB implementation of robust tube-based Model Predictive Control (MPC). The framework provides tools to design and simulate controllers that maintain stability and constraint satisfaction in the presence of bounded disturbances. Tube-based MPC achieves robustness by combining a nominal trajectory planner with an error feedback controller that keeps the actual system state within a "tube" around the nominal trajectory. This repository includes example scripts and implementations demonstrating how to apply the method to control problems. It is particularly useful for researchers, students, and engineers exploring robust control strategies in uncertain environments. By offering a structured implementation, robust-tube-mpc makes it easier to study and extend advanced MPC techniques for real-world applications.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    Rubix ML

    Rubix ML

    A high-level machine learning and deep learning library for PHP

    Rubix ML is a free open-source machine learning (ML) library that allows you to build programs that learn from your data using the PHP language. We provide tools for the entire machine learning life cycle from ETL to training, cross-validation, and production with over 40 supervised and unsupervised learning algorithms. In addition, we provide tutorials and other educational content to help you get started using ML in your projects. Our intuitive interface is quick to grasp while hiding alot of power and complexity. Write less code and iterate faster leaving the hard stuff to us. Rubix ML utilizes a versatile modular architecture that is defined by a few key abstractions and their types and interfaces. Train models in a fraction of the time by installing the optional Tensor extension powered by C. Learners such as neural networks will automatically get a performance boost.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Planfix: Manage Projects, Team's Tasks and Business Processes Icon
    Planfix: Manage Projects, Team's Tasks and Business Processes

    All-in-One Enterprise-Level Software is Now Available for SMB

    Planfix is like a souped-up business process management system for folks who really know their stuff. It's built to help you dive deeper and gives you more options than your run-of-the-mill project and task management systems. Best part? Even small businesses and non-profits can get in on the action.
    Learn More
  • 5
    SHAP

    SHAP

    A game theoretic approach to explain the output of ml models

    SHAP (SHapley Additive exPlanations) is a game theoretic approach to explain the output of any machine learning model. It connects optimal credit allocation with local explanations using the classic Shapley values from game theory and their related extensions. While SHAP can explain the output of any machine learning model, we have developed a high-speed exact algorithm for tree ensemble methods. Fast C++ implementations are supported for XGBoost, LightGBM, CatBoost, scikit-learn and pyspark tree models. To understand how a single feature effects the output of the model we can plot the SHAP value of that feature vs. the value of the feature for all the examples in a dataset. Since SHAP values represent a feature's responsibility for a change in the model output, the plot below represents the change in predicted house price as RM (the average number of rooms per house in an area) changes.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    SageMaker Hugging Face Inference Toolkit

    SageMaker Hugging Face Inference Toolkit

    Library for serving Transformers models on Amazon SageMaker

    SageMaker Hugging Face Inference Toolkit is an open-source library for serving Transformers models on Amazon SageMaker. This library provides default pre-processing, predict and postprocessing for certain Transformers models and tasks. It utilizes the SageMaker Inference Toolkit for starting up the model server, which is responsible for handling inference requests. For the Dockerfiles used for building SageMaker Hugging Face Containers, see AWS Deep Learning Containers. The SageMaker Hugging Face Inference Toolkit implements various additional environment variables to simplify your deployment experience. The Hugging Face Inference Toolkit allows user to override the default methods of the HuggingFaceHandlerService. SageMaker Hugging Face Inference Toolkit is licensed under the Apache 2.0 License.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    TorchIO

    TorchIO

    Medical imaging toolkit for deep learning

    TorchIO is an open-source Python library for efficient loading, preprocessing, augmentation and patch-based sampling of 3D medical images in deep learning, following the design of PyTorch. It includes multiple intensity and spatial transforms for data augmentation and preprocessing. These transforms include typical computer vision operations such as random affine transformations and also domain-specific ones such as simulation of intensity artifacts due to MRI magnetic field inhomogeneity (bias) or k-space motion artifacts. TorchIO is a Python package containing a set of tools to efficiently read, preprocess, sample, augment, and write 3D medical images in deep learning applications written in PyTorch, including intensity and spatial transforms for data augmentation and preprocessing. Transforms include typical computer vision operations such as random affine transformations and also domain-specific ones such as simulation of intensity artifacts due to MRI magnetic field inhomogeneity.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Vearch

    Vearch

    A distributed system for embedding-based vector retrieval

    Vearch is the vector search infrastructure for deep learning and AI applications. Vearch is a distributed vector storage and retrieval system which can be easily extended to billions scale. Vearch implements a high-performance, lockless real-time vector indexing subsystem that utilizes various optimization techniques to support millisecond vector update and retrieval. End-to-end one-click deployment. Through the module of the plugin, a complete default visual search system can be deployed just with one click. Otherwise, you can easily customize your own image, video, or text feature extraction algorithm plugin. This GIF provides a clear demonstration of the project vearch usage and its internal structure. The use of vearch is mainly divided into three steps. Firstly, create DB and Space, then import your data, and finally, you can search on your own dataset.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    VisualDL

    VisualDL

    Deep Learning Visualization Toolkit

    VisualDL, a visualization analysis tool of PaddlePaddle, provides a variety of charts to show the trends of parameters and visualizes model structures, data samples, histograms of tensors, PR curves , ROC curves and high-dimensional data distributions. It enables users to understand the training process and the model structure more clearly and intuitively so as to optimize models efficiently. VisualDL provides various visualization functions, including tracking metrics in real-time, visualizing the model structure, displaying the data sample, visualizing the relationship between hyperparameters and model metrics, presenting the changes of distributions of tensors, showing the pr curves, projecting high-dimensional data to a lower dimensional space and more. Additionally, VisualDL provides VDL.service, which enables developers easily to save, track and share visualization results of experiments.
    Downloads: 3 This Week
    Last Update:
    See Project
  • IT Asset Management (ITAM) Software Icon
    IT Asset Management (ITAM) Software

    Supercharge Your IT Assets, the Easy Way

    EZO AssetSonar is a comprehensive IT asset management platform that provides real-time visibility into your entire digital infrastructure. Track and optimize hardware, software, and license management to reduce risks, control IT spend, and improve compliance.
    Learn More
  • 10
    jieba

    jieba

    Stuttering Chinese word segmentation

    "Jaba" Chinese word segmentation, do the best Python Chinese word segmentation component. Four word segmentation modes are supported. Precise mode, which tries to cut the sentence most precisely, suitable for text analysis. Full mode, scans all the words that can be formed into words in the sentence, the speed is very fast, but the ambiguity cannot be resolved. The search engine mode, on the basis of the precise mode, divides the long words again to improve the recall rate, which is suitable for word segmentation in search engines. The paddle mode uses the PaddlePaddle deep learning framework to train the sequence labeling (bidirectional GRU) network model to achieve word segmentation. Also supports part-of-speech tagging. To use paddle mode, you need to install paddlepaddle-tiny, pip install paddlepaddle-tiny==1.6.1. Currently paddle mode supports jieba v0.40 and above. For versions below jieba v0.40, please upgrade jieba, pip install jieba --upgrade.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    oneDNN

    oneDNN

    oneAPI Deep Neural Network Library (oneDNN)

    This software was previously known as Intel(R) Math Kernel Library for Deep Neural Networks (Intel(R) MKL-DNN) and Deep Neural Network Library (DNNL). oneAPI Deep Neural Network Library (oneDNN) is an open-source cross-platform performance library of basic building blocks for deep learning applications. oneDNN is part of oneAPI. The library is optimized for Intel(R) Architecture Processors, Intel Processor Graphics and Xe Architecture graphics. oneDNN has experimental support for the following architectures: Arm* 64-bit Architecture (AArch64), NVIDIA* GPU, OpenPOWER* Power ISA (PPC64), IBMz* (s390x), and RISC-V. oneDNN is intended for deep learning applications and framework developers interested in improving application performance on Intel CPUs and GPUs. Deep learning practitioners should use one of the applications enabled with oneDNN.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    Audiomentations

    Audiomentations

    A Python library for audio data augmentation

    A Python library for audio data augmentation. Inspired by albumentations. Useful for deep learning. Runs on CPU. Supports mono audio and multichannel audio. Can be integrated in training pipelines in e.g. Tensorflow/Keras or Pytorch. Has helped people get world-class results in Kaggle competitions. Is used by companies making next-generation audio products. Mix in another sound, e.g. a background noise. Useful if your original sound is clean and you want to simulate an environment where background noise is present. A folder of (background noise) sounds to be mixed in must be specified. These sounds should ideally be at least as long as the input sounds to be transformed. Otherwise, the background sound will be repeated, which may sound unnatural. Note that the gain of the added noise is relative to the amount of signal in the input. This implies that if the input is completely silent, no noise will be added.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    Best-of Machine Learning with Python

    Best-of Machine Learning with Python

    A ranked list of awesome machine learning Python libraries

    This curated list contains 900 awesome open-source projects with a total of 3.3M stars grouped into 34 categories. All projects are ranked by a project-quality score, which is calculated based on various metrics automatically collected from GitHub and different package managers. If you like to add or update projects, feel free to open an issue, submit a pull request, or directly edit the projects.yaml. Contributions are very welcome! General-purpose machine learning and deep learning frameworks.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    Datasets

    Datasets

    Hub of ready-to-use datasets for ML models

    Datasets is a library for easily accessing and sharing datasets, and evaluation metrics for Natural Language Processing (NLP), computer vision, and audio tasks. Load a dataset in a single line of code, and use our powerful data processing methods to quickly get your dataset ready for training in a deep learning model. Backed by the Apache Arrow format, process large datasets with zero-copy reads without any memory constraints for optimal speed and efficiency. We also feature a deep integration with the Hugging Face Hub, allowing you to easily load and share a dataset with the wider NLP community. There are currently over 2658 datasets, and more than 34 metrics available. Datasets naturally frees the user from RAM memory limitation, all datasets are memory-mapped using an efficient zero-serialization cost backend (Apache Arrow). Smart caching: never wait for your data to process several times.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    DeepDetect

    DeepDetect

    Deep Learning API and Server in C++14 support for Caffe, PyTorch

    The core idea is to remove the error sources and difficulties of Deep Learning applications by providing a safe haven of commoditized practices, all available as a single core. While the Open Source Deep Learning Server is the core element, with REST API, and multi-platform support that allows training & inference everywhere, the Deep Learning Platform allows higher level management for training neural network models and using them as if they were simple code snippets. Ready for applications of image tagging, object detection, segmentation, OCR, Audio, Video, Text classification, CSV for tabular data and time series. Neural network templates for the most effective architectures for GPU, CPU, and Embedded devices. Training in a few hours and with small data thanks to 25+ pre-trained models. Full Open Source, with an ecosystem of tools (API clients, video, annotation, ...) Fast Server written in pure C++, a single codebase for Cloud, Desktop & Embedded.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    Deeplearning.ai

    Deeplearning.ai

    Study notes, summaries, and auxiliary materials for deep learning

    Deeplearning.ai collects study notes, summaries, and auxiliary materials aligned with the popular deep learning course series many learners take early in their AI journey. It distills core ideas such as optimization, regularization, convolutional networks, sequence models, and practical training tricks. The explanations aim to bridge theory and practice, often connecting mathematical intuition to code-level implications. By organizing the content as “books” or structured notes, it gives students a consistent reference to revisit as models and tooling evolve. Many learners use it to supplement course videos, reinforcing concepts before implementing assignments or projects. As a consolidated guide, it reduces context-switching and helps build a durable mental model of deep learning fundamentals.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    Detect and Track

    Detect and Track

    Code release for "Detect to Track and Track to Detect", ICCV 2017

    Detect-Track is the official implementation of the ICCV 2017 paper Detect to Track and Track to Detect by Christoph Feichtenhofer, Axel Pinz, and Andrew Zisserman. The framework unifies object detection and tracking into a single pipeline, allowing detection to support tracking and tracking to enhance detection performance. Built upon a modified version of R-FCN, the code provides implementations using backbone networks such as ResNet-50, ResNet-101, ResNeXt-101, and Inception-v4, with results demonstrating state-of-the-art accuracy on the ImageNet VID dataset. The repository includes MATLAB-based training and testing scripts, along with pre-trained models and pre-computed region proposals for reproducibility. Multiple testing configurations are available, including multi-frame input and enhanced versions that refine tracking boxes and integrate detection confidence across frames.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    Edward

    Edward

    A probabilistic programming language in TensorFlow

    A library for probabilistic modeling, inference, and criticism. Edward is a Python library for probabilistic modeling, inference, and criticism. It is a testbed for fast experimentation and research with probabilistic models, ranging from classical hierarchical models on small data sets to complex deep probabilistic models on large data sets. Edward fuses three fields, Bayesian statistics and machine learning, deep learning, and probabilistic programming. Edward is built on TensorFlow. It enables features such as computational graphs, distributed training, CPU/GPU integration, automatic differentiation, and visualization with TensorBoard. Expectation-Maximization, pseudo-marginal and ABC methods, and message passing algorithms.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    Image classification models for Keras

    Image classification models for Keras

    Keras code and weights files for popular deep learning models

    All architectures are compatible with both TensorFlow and Theano, and upon instantiation the models will be built according to the image dimension ordering set in your Keras configuration file at ~/.keras/keras.json. For instance, if you have set image_dim_ordering=tf, then any model loaded from this repository will get built according to the TensorFlow dimension ordering convention, "Width-Height-Depth". Pre-trained weights can be automatically loaded upon instantiation (weights='imagenet' argument in model constructor for all image models, weights='msd' for the music tagging model). Weights are automatically downloaded if necessary, and cached locally in ~/.keras/models/. This repository contains code for the following Keras models, VGG16, VGG19, ResNet50, Inception v3, and CRNN for music tagging.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    MegEngine

    MegEngine

    Easy-to-use deep learning framework with 3 key features

    MegEngine is a fast, scalable and easy-to-use deep learning framework with 3 key features. You can represent quantization/dynamic shape/image pre-processing and even derivation in one model. After training, just put everything into your model and inference it on any platform at ease. Speed and precision problems won't bother you anymore due to the same core inside. In training, GPU memory usage could go down to one-third at the cost of only one additional line, which enables the DTR algorithm. Gain the lowest memory usage when inferencing a model by leveraging our unique pushdown memory planner. NOTE: MegEngine now supports Python installation on Linux-64bit/Windows-64bit/MacOS(CPU-Only)-10.14+/Android 7+(CPU-Only) platforms with Python from 3.5 to 3.8. On Windows 10 you can either install the Linux distribution through Windows Subsystem for Linux (WSL) or install the Windows distribution directly. Many other platforms are supported for inference.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 21
    Requests for Research

    Requests for Research

    A living collection of deep learning problems

    Requests for Research is an OpenAI repository that collects and organizes open research ideas in artificial intelligence. It is structured as a curated list of project proposals, challenges, and exploratory directions suggested by OpenAI researchers for the broader community. Each request highlights a specific problem area, often with context, motivation, and possible approaches, serving as inspiration for independent researchers, students, and practitioners. The repository is intended to foster collaboration and accelerate progress by sharing promising questions that OpenAI itself may not have the resources to fully pursue. While the ideas are not guaranteed to be unique or unexplored, they reflect areas OpenAI believes would benefit from more investigation. The project functions as a living document of research directions rather than a codebase, making it an accessible entry point for those looking to contribute to the advancement of AI.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    Scalable Distributed Deep-RL

    Scalable Distributed Deep-RL

    A TensorFlow implementation of Scalable Distributed Deep-RL

    Scalable Agent is the open implementation of IMPALA (Importance Weighted Actor-Learner Architectures), a highly scalable distributed reinforcement learning framework developed by Google DeepMind. IMPALA introduced a new paradigm for efficiently training agents across large-scale environments by decoupling acting and learning processes. In this architecture, multiple actor processes interact with their environments in parallel to collect trajectories, which are then asynchronously sent to a centralized learner for policy updates. The learner uses importance weighting to correct for policy lag between actors and the learner, enabling stable off-policy training at scale. This design allows the system to scale efficiently to hundreds of environments and billions of frames while maintaining sample efficiency and stability. The implementation supports training in DeepMind Lab (DMLab) and has also been adapted for other environments like Atari and Street View.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    SimSiam

    SimSiam

    PyTorch implementation of SimSiam

    SimSiam is a PyTorch implementation of “Exploring Simple Siamese Representation Learning” by Xinlei Chen and Kaiming He. The project introduces a minimalist approach to self-supervised learning that avoids negative pairs, momentum encoders, or large memory banks—key complexities of prior contrastive methods. SimSiam learns image representations by maximizing similarity between two augmented views of the same image through a Siamese neural network with a stop-gradient operation, preventing feature collapse. This elegant yet effective design achieves strong results in unsupervised learning benchmarks such as ImageNet without requiring contrastive losses. The repository provides scripts for both unsupervised pre-training and linear evaluation, using a ResNet-50 backbone by default. It is compatible with multi-GPU distributed training and can be fine-tuned or transferred to downstream tasks like object detection following the same setup as MoCo.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    Techniques

    Techniques

    Techniques for deep learning with satellite & aerial imagery

    This repository is a comprehensive, curated collection of deep learning techniques and best practices specifically applied to satellite and aerial imagery. It covers everything from preprocessing and annotation to model architectures and open datasets. The guide includes code snippets, links to research papers, and hands-on tools, making it valuable for researchers, engineers, and enthusiasts working in remote sensing and geospatial AI.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    Apache MXNet (incubating)

    Apache MXNet (incubating)

    A flexible and efficient library for deep learning

    Apache MXNet is an open source deep learning framework designed for efficient and flexible research prototyping and production. It contains a dynamic dependency scheduler that automatically parallelizes both symbolic and imperative operations. On top of this is a graph optimization layer, overall making MXNet highly efficient yet still portable, lightweight and scalable.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB