Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
Ready-to-use OCR with 80+ supported languages
Single-cell analysis in Python
The official Python client for the Huggingface Hub
FlashInfer: Kernel Library for LLM Serving
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Everything you need to build state-of-the-art foundation models
Uplift modeling and causal inference with machine learning algorithms
Operating LLMs in production
A unified framework for scalable computing
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
Uncover insights, surface problems, monitor, and fine tune your LLM
AIMET is a library that provides advanced quantization and compression
Gaussian processes in TensorFlow
DoWhy is a Python library for causal inference
Easiest and laziest way for building multi-agent LLMs applications
Optimizing inference proxy for LLMs
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Superduper: Integrate AI models and machine learning workflows
Integrate, train and manage any AI models and APIs with your database
The Triton Inference Server provides an optimized cloud
Trainable models and NN optimization tools
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs