AnyScale Ray
Platform for scaling Python and AI applications with Ray.
Curated tools and task pages for Deployment & DevOps. Verified links and trust signals.
Find the best tools for specific deployment & devops tasks.
Platform for scaling Python and AI applications with Ray.
Open-source tool for packaging ML models as Docker containers.
Serverless GPU platform for deploying AI models with auto-scaling.
Kubernetes-native model serving for scalable ML inference.
Open-source inference server for ML models with multi-framework support.
Platform for deploying ML models from notebooks to production APIs.
TensorFlow's production serving system for ML models with batching.
PyTorch's model serving framework for deploying trained models.
Developer platform for provisioning GPU instances for AI work.
European GPU cloud provider for AI training and inference.
GPU cloud optimized for AI model fine-tuning and inference.
Affordable GPU instances for AI development and training.
Serverless AI inference platform for running open-source models.
Distributed AI inference platform for running models across edge nodes.
Platform for enterprise LLM fine-tuning and deployment.
Serverless platform for running data and ML pipelines with GPU support.
Affordable GPU cloud API for running and fine-tuning AI models.
GPU cloud platform for AI inference with model hosting and scaling.
Platform for dispatching Python code to any cloud compute for AI workloads.
Serverless GPU infrastructure for deploying AI models with auto-scaling.
Serverless GPU cloud for deploying AI models with simple Python decorators.
Serverless GPU platform for deploying and serving AI models.
Open-source platform for deploying ML models as production APIs.
Distributed GPU cloud for AI inference with global edge deployments.
Managed API for running open-source AI models without infrastructure.
Kubernetes AI toolchain operator for deploying models on Azure.
Open-source platform for running AI models on Kubernetes clusters.
Developer platform for building and deploying AI applications at scale.
Platform for deploying AI models as production-ready API endpoints.
Efficient AI inference platform for running generative AI models in production.
Cloud platform for deploying GPU-powered AI applications.
Open-source framework for running AI workloads across cloud providers.
Affordable GPU cloud for AI training and inference workloads.
NVIDIA's open-source AI model serving platform for production.
Open-source model serving platform for Kubernetes with autoscaling.
Open-source platform for deploying and monitoring ML models on Kubernetes.
ML deployment platform for shipping models from notebook to production.
GitOps continuous delivery tool for Kubernetes. Provides declarative deployment from Git repositories. Includes progressive delivery and application management. Used for GitOps Kubernetes deployments. Differentiator: Git-based deployment. Integrates with popular business tools and platforms. Supports multiple use cases and deployment scenarios. Includes comprehensive documentation and support resources. Designed for ease of use and quick implementation. Scales from startups to enterprise organizations.
Container platform for packaging and deploying applications. Provides image building, registry, and container runtime. Includes Compose for multi-container applications. Used for containerized deployments. Differentiator: standard container platform. Integrates with popular business tools and platforms. Supports multiple use cases and deployment scenarios. Includes comprehensive documentation and support resources. Designed for ease of use and quick implementation. Scales from startups to enterprise organizations.
Open-source GitOps tool for Kubernetes continuous delivery. Provides Git-based deployment synchronization and automation. Includes notifications and scalability. Used for declarative Kubernetes deployments. Differentiator: GitOps-native. Integrates with popular business tools and platforms. Supports multiple use cases and deployment scenarios. Includes comprehensive documentation and support resources. Designed for ease of use and quick implementation. Scales from startups to enterprise organizations.
Orchestration platform for containers, VMs, and other workloads. Provides unified scheduling across multiple infrastructure types. Includes federation and scaling. Used for heterogeneous workload orchestration. Differentiator: multi-workload support. Integrates with popular business tools and platforms. Supports multiple use cases and deployment scenarios. Includes comprehensive documentation and support resources. Designed for ease of use and quick implementation. Scales from startups to enterprise organizations.
Zero-knowledge environment variable encryption for teams using .env files. Dotenv Vault encrypts .env files with end-to-end encryption. Local development support for all systems. Encrypted syncing across machines and devices. Automatic secret rotation on schedules. Audit logging for tracking access. Integration with deployment platforms. Simple onboarding process. Free and paid tiers available. Best for teams wanting simple secrets management.
Compiler framework for optimizing and deploying machine learning models on diverse hardware. Apache TVM generates efficient code for CPUs, GPUs, TPUs, mobile. Automatic code generation. Memory and compute optimization. Quantization and pruning support. Works with TensorFlow, PyTorch, ONNX. Community-driven. Free and open source. Best for deploying models across hardware targets. Research and production ready.
Cloud gateway service for managing millions of IoT and edge devices. Azure IoT Hub provides device management, monitoring, and secure communication. Supports MQTT, AMQP, HTTPS. Device twin technology for state and metadata. Direct methods for cloud-to-device commands. Twin desired and reported properties. Integration with Azure services. Monitoring and diagnostics. Enterprise security. Pay per million operations. Best for Microsoft ecosystem. Handles billions of messages.
Container-based platform for deploying applications to IoT edge devices and fleets. Balena handles device management, updates, and monitoring at scale. Supports Raspberry Pi, Jetson, x86 hardware. Docker containers for application deployment. Over-the-air updates. Offline capability. Local mode for development. Balena Cloud dashboard. Git-based workflows. Pay per active device. Open-source balena Engine. Best for device fleet management.
Open-source network video recorder with real-time AI object detection. Frigate runs on standard hardware with GPU acceleration. Real-time person, car, dog, cat detection using YOLO. Event-based recording saves storage. Mobile app for viewing. REST API. Integration with Home Assistant. Supports multiple camera streams. Privacy-focused local processing. Free and open source. Best for home and small business security.
Fully managed IoT platform for connecting, processing, and managing IoT devices at scale. Google Cloud IoT Engine handles secure device communication, authentication, and data ingest. Integrates with Cloud Pub/Sub for streaming. Cloud Dataflow for processing. Firestore for device state. MQTT and HTTP support. Device manager for fleet management. Monitoring and logging. Pay per connection hour. Best for Google Cloud users. Handles millions of concurrent devices.
Toolkit for deploying computer vision and AI models to edge devices optimized for Intel hardware. OpenVINO provides model optimization, inference engine, and tools. Supports TensorFlow, PyTorch, ONNX. Model converter for optimization. Faster inference and lower memory. Works across Intel CPUs, GPUs, VPUs. Scales from data center to edge. Free and open source. Best for deploying vision models to edge devices.
AI computing platform for edge devices and robotics with powerful GPUs. Jetson provides hardware modules and software stack for AI inference at the edge. Supports CUDA, TensorRT for optimized models. Multiple SKUs from small Nano to Xavier. JetPack SDK with Linux OS. Container support. Power efficient. Real-time performance. Best for computer vision and AI-intensive edge workloads. Used in robots and autonomous systems.
Open-source inference engine for running ONNX models across devices and platforms. ONNX Runtime supports CPUs, GPUs, TPUs, and specialized accelerators. Cross-platform support for Windows, Linux, macOS, mobile. Optimizations for model inference. Quantization and pruning support. C++, C#, Python, Java APIs. Works on edge devices. Community-driven. Free and open source. Best for model interoperability.