When creating an inference service, users can now select either "Standard Mode" or "Advanced Mode," with "Standard Mode" being the default option.
A new "Dashboards" feature has been introduced, allowing users to customize and add dashboard charts according to their needs. For example, projects using GPUs from vendors other than NVIDIA can add customized dashboards provided by the manufacturers.
The new "Alauda AI Workbench" plugin is available for installation, providing users with IDE environments such as Jupyter Notebook and VS Code. This plugin replaces the "advanced" capabilities of the previous version and streamlines unnecessary components and some functions originally found in Kubeflow.
A native Kubeflow solution has been launched to meet the needs of project clients who are accustomed to using the native capabilities of the Kubeflow community.
A multi-node multi-GPU solution has been introduced to cater to users' requirements for deploying models with large parameter counts.
Notebook-based solutions for model pre-training and fine-tuning have been launched to support users in optimizing their models.
An inference service authentication solution based on Enovy AI Gateway has been introduced, supporting the creation of API Keys for inference services to enhance permission control capabilities.
The logging functionality for inference services has been enhanced, including features such as automatic log updates, pause updates, and container switching, maintaining consistency with Alauda Container Platform capabilities.
Due to limited use cases and its impact on the user experience of large model inference services, the mlserver inference runtime has been downgraded to a solution. It is no longer included in the product by default, but a solution is provided to support specific scenarios, such as Small Language Model inference.
Both the Apps feature and Dify are positioned as AI Agent development capabilities, with Dify offering a simpler development approach through its low-code capabilities. In contrast, the pure customization and from-scratch development approach of the Apps feature is less convenient. Therefore, the Apps feature has been discontinued. Projects that require pure custom development of AI Agents can be accommodated through alternative solutions.
There are two ways to upload models: via git push command-line or through the UI. Command-line uploads offer better performance and faster speeds. Although the UI upload is user-friendly, it tends to freeze when dealing with large model files, which are typically several hundred GB in size. Therefore, the UI upload feature has been discontinued. To facilitate user access, a documentation link has been added in place of the original feature, allowing users to quickly navigate to the user manual for operation commands.
No issues in this release.