Upgrade
This article describes how to upgrade from GPU-manager or an older Hami version to the latest Hami release.
GPU-manager to Hami
Note
- GPU-manager and Hami cannot be deployed on the same node, but they can coexist in the same cluster.
- During the upgrade, applications must be modified one by one, which will restart the corresponding workload pods.
- When you only have one GPU node, you must uninstall GPU-manager before installing Hami. You can achieve this by switching the node label while the two plugins are deployed.
For example, remove the
nvidia-device-enable=vgpunode label to delete the GPU-manager instance on that node, then add thegpu=onlabel to deploy the Hami plugin on it.
Procedure
Modify your applications one by one. Example:
Your old GPU-manager instance:
Migrate to Hami:
Hami to Hami
Important Changes (v2.5.0 → v2.8.0)
⚠️ Upgrading from v2.5 to v2.8.0 should not affect existing applications. ✅ It is recommended to restart applications with a rolling update to avoid unexpected issues.
Procedure
- Upgrade ACP version if needed.
- Upload the package of Hami v2.8.0 plugin to ACP.
- Go to the
Administrator->Clusters->Target Cluster->Functional Componentspage, then click theUpgradebutton and you will see theAlauda Build of HAMican be upgraded. - Update some ConfigMaps that defines extended resources, which can be used to set extended resources on the ACP. Run the following script in your gpu cluster:
Click to expand code
Note
If you configured resource quota for HAMi resources in versions prior to v2.7.1, please delete and reconfigure it.
If you are upgrading to Hami v2.8.0 and also use Alauda Build of Hami-WebUI, make sure the WebUI version is v1.10.0. Earlier v1.5.0 is not compatible with Hami v2.8.