Step 5: Deploy a model
Model deployments make trained models available in your environment. Follow these steps to create, activate, and verify model deployments.
Create a deployment
To deploy a model, specify its modelId
, deployment name, spaceId
, and other configuration details in the
request.
You also need to set enablePredictions
to true
to enable predictions for this model in the following API call:
curl -L -X POST "https://<TENANT>/api/v1/ml/deployments" ^
-H "Authorization: Bearer <ACCESS_TOKEN>" ^
-H "Content-Type: application/json" ^
-d "{
\"data\": {
\"type\": \"deployment\",
\"attributes\": {
\"name\": \"<NAME>\",
\"spaceId\": \"<SPACE_ID>\",
\"description\": \"<DESCRIPTION>\",
\"modelId\": \"<MODEL_ID>\",
\"enablePredictions\": true,
\"deprecated\": false
}
}
}"
If the deployment is successful, the 201
response includes:
- the
id
, which is the deployment ID that you’ll need for activating the model and generating predictions. - the
modelId
, which is the model ID associated with the deployment.
Response example
{
"data": {
"type": "deployment",
"id": "4215f16e-4cc7-432c-bb5d-fc02f16b9c06",
"attributes": {
"id": "4215f16e-4cc7-432c-bb5d-fc02f16b9c06",
"createdAt": "2024-12-01T16:34:31.552381459Z",
"updatedAt": "2024-12-01T16:34:31.552381459Z",
"name": "Model deployment for churn predictions",
"spaceId": "6745f737f536738170dfe82f",
"description": "Model deployment for churn predictions",
"modelId": "bed12707-7deb-4a83-99f1-c936aa4a0ebd",
"enablePredictions": true,
"deprecated": false,
"createdBy": "67475097984561d02f0cb3dc",
"ownerId": "67475097984561d02f0cb3dc",
"errorMessage": null
}
}
}
With your model deployed in your environment, you can now activate it to make it ready to generate predictions.
Activate a model
Once a deployment is created, it needs to be activated to generate predictions. Use the deployment ID from the previous step to activate the model.
You can activate a model with the following API call:
curl -L -X POST "https://<TENANT>/api/v1/ml/deployments/<DEPLOYMENT_ID>/actions/activate-models" ^
-H "Authorization: Bearer <ACCESS_TOKEN>"
When the model is successfully activated, you’ll get a 204
response.
Verify model activation
After activation, verify that the model’s status is updated to enabled
.
Use the model’s modelId
to review its details with the following API call:
curl -L "https://<TENANT>/api/v1/ml/experiments/<EXPERIMENT_ID>/models/<MODEL_ID>" ^
-H "Authorization: Bearer <ACCESS_TOKEN>"
In the response, "modelState": "enabled"
means that the model is activated and ready to generate predictions.
Next step
With the model deployment activated, you can now generate predictions in real-time or batch. Predictions allow you to use the deployed model for inference on new data.