Skip to main content

Overview

OGC Serverless Kubernetes is the simplest and quickest way to run AI/ML workloads on Kubernetes. It enables developers to focus on the workloads while the nodes are automatically managed.

  • GPU Support: Integrated GPU resources make it ideal for compute-intensive applications, particularly in ML and AI.
  • Serverless Managed Cluster: The Kubernetes environment is fully managed by Ori, removing the overhead of managing clusters and node resources.
  • Familiar Experience: Offers a user experience similar to traditional Kubernetes, providing familiarity for those accustomed to kubectl.
  • Cost-Efficient: Per-minute billing ensures users only pay for the resources they consume, optimizing cost efficiency.
  • Scalability: Dynamically scales based on workload demands, ensuring optimal resource utilization.

Currently, we are offering the following GPU types:

  • NVIDIA H100 SXM: for complex deep learning tasks and high performance inference.
  • NVIDIA H100 PCIe: ideal for inference/training of large models and best performance.
  • NVIDIA L4: great for graphics and for machine learning tasks such as training and inference at a great price point.
  • NVIDIA L40S: designed for high-performance computing, AI workloads, and advanced graphics rendering in DC and Enterprise environments.
info

Serverless Kubernetes is currently in Beta mode.

Use Cases

  • Ideal for applications that require rapid scaling.
  • Suitable for ML/AI workloads that need GPU resources.
  • Perfect for developers who prefer a Kubernetes environment without the complexity of cluster management.

Resources