Import a Model Registry using Triton Server
Model Preparation
- Since our AI Platform only accesses models from a Network Volume, you must first create a Network Volume. Pull your model from local file systems or cloud storage (AWS S3, Azure Blob, or GCS) into the Network Volume.
- Ensure the model is in Triton-compatible format such as:
- ONNX (
.onnx
) - TensorFlow (
SavedModel
or .pb
) - PyTorch TorchScript (
.pt
) - TensorRT (
.engine
) - OpenVINO (
.xml
and .bin
) - Ensemble Model (combining multiple models)
Step 1: Accessing the Model Registry
- Log in to your GreenNode AI Platform account and navigate to the Model Registry Dashboard.
- Find and click on the "Import a model registry" button.
Step 2: Import a Model Registry
- Location & Model registry name: Select the location & a specific name for this model.
- Container: Select the Pre-built container option to use as a supported framework.
- Framework: Choose a model deployment framework & suitable version that meets your requirements. In this tutorial, we select Triton 22.12
- Model source: Access to model stored on network volume: Select network volume as a data mount method for the training job.
Model repository: Specify the location where your model's registry is stored. To use Triton, we need to build a model repository. The structure of the repository as follows:
network-volume
|
+-- model_repository
|
+-- resnet
|
+-- config.pbtxt
+-- 1
|
+-- model.onnx
It should be added to the "network-volume" section of the location path, e.g., "/network-volume/model_repository".
Check the Triton documentation for compatibility guidelines and any necessary adjustments to your model format or configuration.
- Click the “Import” button to complete the process.
Related Articles
Import a Model Registry using vLLM
Step 1: Accessing the Model Registry Log in to your GreenNode AI Platform account and navigate to the Model Registry Dashboard. Find and click on the "Import a model registry" button. Step 2: Import a Model Registry Location & Model registry name: ...
Import a model registry with pre-built container
Model Preparation Ensure your machine learning model is packaged and stored in a container image compatible with Triton Inference Server. Upload the container image containing your model to a storage location accessible by our AI Platform. The online ...
Import a Model Registry using NVIDIA NIM
Step 1: Accessing the Model Registry Log in to your GreenNode AI Platform account and navigate to the Model Registry Dashboard Model Registry Dashboard at: . Find and click on the "Import a model registry" button. Step 2: Import a Model Registry ...
Import a Model Registry
The model registry is a centralized repository for storing trained models, their metadata, versions, and associated artifacts. It allows for version control, management, and organization of models developed during the training phase. This enables ...
Import a model registry with custom container
The model registry is a centralized repository for storing trained models, their metadata, versions, and associated artifacts. It allows for version control, management, and organization of models developed during the training phase. This enables ...