logo
Alauda AI
English
简体中文
English
简体中文
logo
Alauda AI
Navigation

Overview

intro
Features Overview
Quick Start
Release Notes

Install

Pre-installation Configuration
Install Alauda AI Essentials
Install Alauda AI

Upgrade

Upgrade from AML 1.2

Infrastructure Management

Device Management

About Alauda Build of Hami
About Alauda Build of NVIDIA GPU Device Plugin

Multi-Tenant

Guides

Namespace Management
Permissions

Model Deployment & Inference

Overview

Introduction
Features

Inference Service

Introduction

Guides

Feature Introduction

Troubleshooting

Experiencing Inference Service Timeouts with MLServer Runtime
Permissions

how_to

Configure External Access for Inference Services

Model Management

Introduction

Guides

Feature Introduction
Permissions

Monitoring & Ops

Overview

Introduction
Features Overview

Logging & Tracing

Introduction

Guides

Logging

Resource Monitoring

Introduction

Guides

Resource Monitoring
Permissions

API Reference

Introduction

Kubernetes APIs

Manage APIs

AmlNamespace [manage.aml.dev/v1alpha1]

Operator APIs

AmlCluster [amlclusters.aml.dev/v1alpha1]

Inference Service APIs

ClusterServingRuntime [serving.kserve.io/v1alpha1]
InferenceService [serving.kserve.io/v1beta1]
glossary

#Model Deployment & Inference

Overview

Introduction

  • Model Management
  • Inference Service

Features

  • Model Management
  • Inference Service

Inference Service

Introduction

  • Core Advantages
  • Application Scenarios

Guides

    Troubleshooting

      Permissions

        how_to

          Model Management

          Introduction

          • Core Advantages
          • Application Scenarios

          Guides

            Permissions

              📝 Edit this page on GitHub
              Previous PagePermissions
              Next PageOverview