Ensure your machine learning model is packaged and stored in a container image compatible with Triton Inference Server.
Upload the container image containing your model to a storage location accessible by our AI Platform.
The online prediction feature of our AI Platform currently only supports models built using Triton Inference Server. Ensure that your custom container image is compatible with Triton before deploying your model for online prediction.
Log in to your Greennode AI Platform account and navigate to the Model Registry Dashboard at: https://aiplatform.console.greennode.ai/registry.
Find and click on the "Import a model registry" button.
Location & Model registry name: Select the location & a specific name for this model.
Container: Select the Pre-built container option to use as a supported framework.
Model framework & version: Choose a model training framework & suitable version that meets your requirements.
Access to training model stored on network volume: Select network volume as a data mount method for the training job.
Model repository: Specify the location where your model's registry is stored. It should be added to the "network-volume" section of the location path, e.g., "/network-volume/training-model".
Network volume: Select a network volume that you want to access from the training job. This network volume will be mounted at pre-defined folder.
Click the “Import” button to complete the process.
To simplify model deployment, our Model Registry offers pre-built containers. However, please be aware of the following limitations when using the pre-built container option:
Triton Framework Only: The pre-built container exclusively supports the Triton Inference Server framework. If your model is not compatible with Triton, you will need to use a custom container.
Triton Version 22.12: The pre-built container is configured with Triton version 22.12. Your model must be compatible with this specific version to ensure successful deployment.
To deploy your model using the pre-built container, ensure your model registry meets the following criteria:
Triton Compatibility: Your model must be compatible with the Triton Inference Server framework. This typically involves exporting your model in a format that Triton can understand (e.g., ONNX, TensorFlow SavedModel).
Triton Version 22.12 Compatibility: Your model must be compatible with Triton version 22.12. Check the Triton documentation for compatibility guidelines and any necessary adjustments to your model format or configuration.
Model Configuration: Provide a valid config.pbtxt
file that describes your model's configuration for Triton. This file should include details like model name, input and output tensors, and backend settings.
Dependencies: If your model has any specific dependencies (libraries, packages), ensure they are included in your model repository or can be installed within the pre-built container environment.
We are actively working to expand framework and version support for pre-built containers. Your feedback is valuable in helping us prioritize these enhancements.