Overview
OGC Serverless Kubernetes is the simplest and quickest way to run AI/ML workloads on Kubernetes. It enables developers to focus on the workloads while the nodes are automatically managed.
- GPU Support: Integrated GPU resources make it ideal for compute-intensive applications, particularly in ML and AI.
- Serverless Managed Cluster: The Kubernetes environment is fully managed by Ori, removing the overhead of managing clusters and node resources.
- Familiar Experience: Offers a user experience similar to traditional Kubernetes, providing familiarity for those accustomed to kubectl.
- Cost-Efficient: Per-minute billing ensures users only pay for the resources they consume, optimizing cost efficiency.
- Scalability: Dynamically scales based on workload demands, ensuring optimal resource utilization.
Currently, we are offering the following GPU types:
- NVIDIA H100 SXM: for complex deep learning tasks and high performance inference.
- NVIDIA H100 PCIe: ideal for inference/training of large models and best performance.
- NVIDIA L4: great for graphics and for machine learning tasks such as training and inference at a great price point.
- NVIDIA L40S: designed for high-performance computing, AI workloads, and advanced graphics rendering in DC and Enterprise environments.
info
Serverless Kubernetes is currently in Beta mode.
Use Cases
- Ideal for applications that require rapid scaling.
- Suitable for ML/AI workloads that need GPU resources.
- Perfect for developers who prefer a Kubernetes environment without the complexity of cluster management.
Resources
- Go through the Get Started guide
- Learn about how to use Node Selectors to allocate specific GPU types.
- Deploy a Stable Diffusion model into your cluster.
- Understand how billing is done with Kubernetes.