Alauda AI now offers flexible deployment options. Starting with Alauda AI 1.4, the Serverless capability is an optional feature, allowing for a more streamlined installation if it's not needed.
To begin, you will need to deploy the Alauda AI Operator. This is the core engine for all Alauda AI products. By default, it uses the KServe Raw Deployment mode for the inference backend, which is particularly recommended for resource-intensive generative workloads. This mode provides a straightforward way to deploy models and offers robust, customizable deployment capabilities by leveraging foundational Kubernetes functionalities.
If your use case requires Serverless functionality, which enables advanced features like scaling to zero on demand for cost optimization, you can optionally install the Alauda AI Model Serving Operator. This operator is not part of the default installation and can be added at any time to enable Serverless functionality.
Recommended deployment option: For generative inference workloads, the Raw Kubernetes Deployment approach is recommended as it provides the most control over resource allocation and scaling.
Operator Components:
Alauda AI Operator
Alauda AI Operator is the main engine that powers Alauda AI products. It focuses on two core functions: model management and inference services, and provides a flexible framework that can be easily expanded.
Download package: aml-operator.xxx.tgz
Alauda AI Model Serving Operator
Alauda AI Model Serving Operator provides serverless model inference.
Download package: kserveless-operator.xxx.tgz
You can download the app named 'Alauda AI' and 'Alauda AI Model Serving' from the Marketplace on the Customer Portal website.
We need to upload both Alauda AI and Alauda AI Model Serving to the cluster where Alauda AI is to be used.
First, we need to download the violet tool if not present on the machine.
Log into the Web Console and switch to the Administrator view:
violet tool.chmod +x ${PATH_TO_THE_VIOLET_TOOL} to make the tool executable.Save the following script in uploading-ai-cluster-packages.sh first, then read the comments below to update environment variables for configuration in that script.
${PLATFORM_ADDRESS} is your ACP platform address.${PLATFORM_ADMIN_USER} is the username of the ACP platform admin.${PLATFORM_ADMIN_PASSWORD} is the password of the ACP platform admin.${CLUSTER} is the name of the cluster to install the Alauda AI components into.${AI_CLUSTER_OPERATOR_NAME} is the path to the Alauda AI Cluster Operator package tarball.${KSERVELESS_OPERATOR_PKG_NAME} is the path to the KServeless Operator package tarball.${REGISTRY_ADDRESS} is the address of the external registry.${REGISTRY_USERNAME} is the username of the external registry.${REGISTRY_PASSWORD} is the password of the external registry.After configuration, execute the script file using bash ./uploading-ai-cluster-packages.sh to upload both Alauda AI and Alauda AI Model Serving operator.
In Administrator view:
Click Marketplace / OperatorHub.
At the top of the console, from the Cluster dropdown list, select the destination cluster where you want to install Alauda AI.
Select Alauda AI, then click Install.
Install Alauda AI window will popup.
Then in the Install Alauda AI window.
Leave Channel unchanged.
Check whether the Version matches the Alauda AI version you want to install.
Leave Installation Location unchanged, it should be aml-operator by default.
Select Manual for Upgrade Strategy.
Click Install.
Confirm that the Alauda AI tile shows one of the following states:
Installing: installation is in progress; wait for this to change to Installed.Installed: installation is complete.Once Alauda AI Operator is installed, you can create an Alauda AI instance.
In Administrator view:
Click Marketplace / OperatorHub.
At the top of the console, from the Cluster dropdown list, select the destination cluster where you want to install the Alauda AI Operator.
Select Alauda AI, then Click.
In the Alauda AI page, click All Instances from the tab.
Click Create.
Select Instance Type window will pop up.
Locate the AmlCluster tile in Select Instance Type window, then click Create.
Create AmlCluster form will show up.
Keep default unchanged for Name.
Select Deploy Flavor from dropdown:
single-node for non HA deployments.ha-cluster for HA cluster deployments (Recommended for production).Set KServe Mode to Managed.
Input a valid domain for Domain field.
This domain is used by ingress gateway for exposing model serving services. Most likely, you will want to use a wildcard name, like *.example.com.
You can specify the following certificate types by updating the Domain Certificate Type field:
ProvidedSelfSignedACPDefaultIngressBy default, the configuration uses SelfSigned certificate type for securing ingress traffic to your cluster, the certificate is
stored in the knative-serving-cert secret that is specified in the Domain Certificate Secret field.
To use certificate provided by your own, store the certificate secret in the istio-system namespace, then update the value of the
Domain Certificate Secret field, and change the value of the Domain Certificate Secret field to Provided.
In the Serverless Configuration section, set Knative Serving Provider to Operator; leave all other parameters blank.
Under Gitlab section:
cpaas-system for Admin Token Secret Namespace.aml-gitlab-admin-token for Admin Token Secret Name.Review above configurations and then click Create.
Check the status field from the AmlCluster resource which named default:
Should returns Ready:
Now, the core capabilities of Alauda AI have been successfully deployed. If you want to quickly experience the product, please refer to the Quick Start.
Serverless functionality is an optional capability that requires an additional operator and instance to be deployed.
The Serverless capability relies on the Istio Gateway for its networking. Please install the Service Mesh first by following the documentation.
In Administrator view:
Click Marketplace / OperatorHub.
At the top of the console, from the Cluster dropdown list, select the destination cluster where you want to install.
Select Alauda AI Model Serving, then click Install.
Install Alauda AI Model Serving window will popup.
Then in the Install Alauda AI Model Serving window.
Leave Channel unchanged.
Check whether the Version matches the Alauda AI Model Serving version you want to install.
Leave Installation Location unchanged, it should be kserveless-operator by default.
Select Manual for Upgrade Strategy.
Click Install.
Confirm that the Alauda AI Model Serving tile shows one of the following states:
Installing: installation is in progress; wait for this to change to Installed.Installed: installation is complete.Once Alauda AI Model Serving Operator is installed, you can create an instance. There are two ways to do this:
You can have the instance automatically created and managed by the AmlCluster by editing its parameters.
In Administrator view:
Click Marketplace / OperatorHub.
At the top of the console, from the Cluster  dropdown list, select the destination cluster where you previously installed the AmlCluster.
Select Alauda AI, then Click.
In the Alauda AI page, click All Instances from the tab.
Click name default.
Locate Actions dropdown list and select update.
update default form will show up.
In the Serverless Configuration section:
Legacy.Managed.Leave all other parameters unchanged. Click Update.
You can manually create the KnativeServing (knativeservings.components.aml.dev) instance.
In Administrator view:
Click Marketplace / OperatorHub.
At the top of the console, from the Cluster dropdown list, select the destination cluster where you want to install.
Select Alauda AI Model Serving, then Click.
In the Alauda AI Model Serving page, click All Instances from the tab.
Click Create.
Select Instance Type window will pop up.
Locate the KnativeServing tile in Select Instance Type window, then click Create.
Create KnativeServing form will show up.
Keep default-knative-serving unchanged for Name.
Keep knative-serving unchanged for Knative Serving Namespace.
In the Ingress Gateway section, configure the following:
1-22).For details on configuring the domain and certificate type, refer to the relevant section.
In the Values section, configure the following:
Select Deploy Flavor from dropdown:
single-node for non HA deployments.ha-cluster for HA cluster deployments (Recommended for production).Set Global Registry Address to Match Your Cluster
You can find your cluster's private registry address by following these steps:
Private Registry address value in the Basic Info section.Configure the AmlCluster instance to integrate with a KnativeServing instance.
In the AmlCluster instance update window, you will need to fill in the required parameters in the Serverless Configuration section.
After the initial installation, you will find that only the Knative Serving Provider is set to Operator. You will now need to provide values for the following parameters:
components.aml.dev/v1alpha1KnativeServingdefault-knative-servingIf you want to replace GitLab Service after installation, follow these steps:
Reconfigure GitLab Service
Refer to the Pre-installation Configuration and re-execute its steps.
Update Alauda AI Instance
Modify GitLab Configuration
In the Update default form:
cpaas-systemaml-gitlab-admin-tokenRestart Components
Restart the aml-controller deployment in the kubeflow namespace.
Refresh Platform Data
In Alauda AI management view, re-manage all namespaces.
Original models won't migrate automatically Continue using these models: