Skip to main content
CNCF: SandboxLatest Release: v2.8.0

Kubernetes virtualization middleware for heterogeneous AI infrastructure

HAMi enables sharing, isolation and scheduling for GPU/NPU/MLU resources so mixed accelerators run efficiently on one platform.

Runs on Kubernetes
CNCF logoCNCF logo
CNCF Sandbox Project

HAMi is a CNCF Sandbox project

The project is developed in the open under CNCF governance, with contributions from a growing global community of companies and individuals.

Why HAMi

Unified multiplexing across heterogeneous devices

Use one Kubernetes-native workflow to schedule GPU, NPU, MLU and other AI accelerators.

Fine-grained slicing and hard isolation

Allocate memory/core slices precisely for training and inference jobs in mixed workloads, with hard isolation enforced at runtime.

Dynamic controls and scheduling

Supports binpack, spread, node-topology-aware, and task-topology-aware scheduling policies to optimize resource utilization and placement.

Aligned with Kubernetes standards (DRA/CDI)

Build on standard interfaces to avoid lock-in and simplify long-term platform evolution.

Key Features

Kubernetes Native

Zero-change adoption path with Kubernetes-compatible APIs and deployment model.

Open and Vendor Neutral

Community-driven governance and hardware ecosystem support for diverse environments.

Resource Isolation & QoS

Control memory/core usage to improve fairness, reliability and utilization.

Unified Monitoring

Provide consistent metrics and operational visibility across device vendors.

Architecture & How It Works

HAMi works through two core paths: GPU virtualization/slicing and heterogeneous scheduling from request to isolated execution.

View full architecture docs →

Before vs After HAMi

The same Pod requests enter whole-GPU allocation on the left and HAMi slicing on the right, showing how scheduling semantics change placement.

Same workload inputStep 3: place Pod C
Pod A0.3 GPU slicemem 30% · core 25%
Pod B0.25 GPU slicemem 24% · core 20%
Pod C0.2 GPU slicemem 18% · core 15%
Whole-GPU allocation

Without HAMi

Exclusive semantics

Each Pod is scheduled with whole-GPU semantics, so the unused portion of that card cannot be shared with another Pod.

Pod C

Pod C claims an entire GPU, so the remaining capacity on that card becomes stranded.

GPUs consumed3/3
Fragmentation36 cells stranded by exclusive GPU claims
GPU 0
exclusive allocation
31%
Pod A
GPU 1
exclusive allocation
25%
Pod B
GPU 2
exclusive allocation
14%
Pod C
Fine-grained slicing & policy scheduling

With HAMi

Shared slices

The same Pod requests are sliced first, then placed by policy to pack, spread, or respect topology locality.

Pod C

Pod C slice and pack onto the most loaded compatible gpu.

Active GPUs1/3
Policy result36 cells still schedulable for later packing
GPU 0
Topo A
shared slices
70%
Pod APod BPod C
GPU 1
Topo B
open capacity
0%
Available
GPU 2
Topo C
open capacity
0%
Available

Ecosystem & Device Support

Broad accelerator ecosystem across vendors. See docs for full support matrix.

  • Huawei Ascend
  • Cambricon
  • Enflame
  • Hygon
  • Iluvatar
  • Kunlunxin
  • Metax
  • Mthreads
  • NVIDIA
View full supported devices list →

Adopters

The organizations below are evaluating or using HAMi in production environments.

  • 4paradigm
  • Baidu AI Cloud
  • Baihai
  • BONC
  • C. Y. Intell
  • Chaintech
  • China Eastcom
  • China Merchants Bank
  • China Mobile
  • China Unicom Industrial Internet
  • China University of Mining and Technology
  • Coocaa Technology
  • CoresHub
  • DaoCloud
  • DeepRoute.ai
  • Dialo
  • Donghua University
  • ECloud (China Mobile Cloud)
  • Empathy
  • Ghostcloud
  • Gsafety
  • Guangdong University of Technology
  • Guangzhou Pingao
  • H3C
  • Hangzhou Lianhui
  • Haofang
  • Harbin Institute of Technology
  • Huawei
  • i-tudou
  • iFlytek
  • Infervision
  • Institute of Information Engineering, CAS
  • Ke Holdings Inc.
  • Kylinsoft
  • LinkedIn
  • Linklogis
  • Mashang Consumer Finance
  • Miaoyun
  • Nankai University - Network Laboratory
  • NIO
  • Northsoft
  • Ping An Bank
  • Ping An Securities
  • PPIO
  • Prep EDU
  • RageHealth
  • RiseUnion
  • SAP
  • SF Technology
  • Shanghai Ashermed Medical Technology Co., Ltd.
  • si-tech
  • Sina Weibo
  • Sinochem Modern Agriculture
  • Snow
  • Southeast University
  • SZZT
  • Technical University of Munich
  • Tongcheng Travel
  • UCloud
  • Unicdata
  • Viettel
  • Woqu
  • Xuanyuan Network Technology Co., Ltd.
  • XW Bank
  • ZStack

Contributors

HAMi is advanced by contributors from the community and industry. These organizations actively participate in project development and ecosystem collaboration.

  • Dynamia
  • pd4
  • DaoCloud
  • icbc
  • CAIH
  • CNCR
  • XW
  • iflytek
  • huawei
  • qxzg
  • riseunion
  • ascend
  • cambricon
  • enflame
  • hygon
  • iluvatar
  • metax
  • mthreads
CNCFHAMi is a CNCF Sandbox project