Te HAMi community is proud to announce the official release of HAMi v2.8.0. This represents a milestone version in terms of architectural completeness, scheduling reliability, and ecosystem alignment.
v2.8.0 not only introduces multiple key feature updates but also delivers systematic enhancements in Kubernetes native standard alignment, heterogeneous device support, production readiness, and observability, making HAMi more suitable for AI production clusters that require long-term operation with high stability and clear evolution paths.
This article provides a detailed overview of the major updates in v2.8.0.
During the use of HAMi, it is common for Pods to be created and remain in a Pending state, particularly due to the following two issues:
- Pod UnexpectedAdmissionError
- Pod Pending
This section provides a rough walkthrough of the related code to explain the interactions between components during scheduling and how resources are calculated. Other details may be omitted.
What is HAMi?​
HAMi (Heterogeneous AI Computing Virtualization Middleware), formerly known as k8s-vGPU-scheduler, is an innovative solution designed to manage heterogeneous AI computing devices within Kubernetes clusters. This all-in-one middleware enables the sharing of various AI devices while ensuring resource isolation among different tasks. By improving the utilization rates of heterogeneous computing devices, HAMi provides a unified multiplexing interface that caters to diverse device types.