logo
Alauda AI
English
简体中文
English
简体中文
logo
Alauda AI
Navigation

Overview

intro
Features Overview
Quick Start
Release Notes

Install

Pre-installation Configuration
Install Alauda AI Essentials
Install Alauda AI

Upgrade

Upgrade from AML 1.2

Infrastructure Management

Device Management

About Alauda Build of Hami
About Alauda Build of NVIDIA GPU Device Plugin

Multi-Tenant

Guides

Namespace Management
Permissions

Model Deployment & Inference

Overview

Introduction
Features

Inference Service

Introduction

Guides

Feature Introduction

Troubleshooting

Experiencing Inference Service Timeouts with MLServer Runtime
Permissions

how_to

Configure External Access for Inference Services

Model Management

Introduction

Guides

Feature Introduction
Permissions

Monitoring & Ops

Overview

Introduction
Features Overview

Logging & Tracing

Introduction

Guides

Logging

Resource Monitoring

Introduction

Guides

Resource Monitoring
Permissions

API Reference

Introduction

Kubernetes APIs

Manage APIs

AmlNamespace [manage.aml.dev/v1alpha1]

Operator APIs

AmlCluster [amlclusters.aml.dev/v1alpha1]

Inference Service APIs

ClusterServingRuntime [serving.kserve.io/v1alpha1]
InferenceService [serving.kserve.io/v1beta1]
glossary

#Guides

Feature Introduction

  • Advantages
  • Applicable Scenarios
  • Value Brought
  • Main Features
  • Technical Notes
  • Create Model Repository
📝 Edit this page on GitHub
Previous PageIntroduction
Next PageFeature Introduction