Predictor#
You can get a ModelServing instance using Project.get_model_serving. Once you have it, you can create a predictor using ModelServing.create_predictor. Predictors can also be accessed from the Deployment metadata objects:
deployment.predictor
Predictor #
Bases: DeployableComponent
Metadata object representing a predictor in Model Serving.
api_protocol property writable #
api_protocol
API protocol enabled in the predictor (e.g., HTTP or GRPC).
artifact_files_path property #
artifact_files_path
Path of the artifact files deployed by the predictor.
artifact_path property #
artifact_path
Path of the model artifact deployed by the predictor. Resolves to /Projects/{project_name}/Models/{name}/{version}/Artifacts/{artifact_version}/{name}{version}.zip.
artifact_version property writable #
artifact_version
Artifact version deployed by the predictor.
Deprecated
Artifact versions are deprecated in favor of deployment versions.
config_file property writable #
config_file
Model server configuration file passed to the model deployment.
It can be accessed via CONFIG_FILE_PATH environment variable from a predictor or transformer script. For LLM deployments without a predictor script, this file is used to configure the vLLM engine.
inference_logger property writable #
inference_logger
Configuration of the inference logger attached to this predictor.
model_framework property writable #
model_framework
Model framework of the model to be deployed by the predictor.
requested_instances property #
requested_instances
Total number of requested instances in the predictor.
deploy #
deploy()
Create a deployment for this predictor and persists it in the Model Serving.
Example
import hopsworks
project = hopsworks.login()
# get Hopsworks Model Registry handle
mr = project.get_model_registry()
# retrieve the trained model you want to deploy
my_model = mr.get_model("my_model", version=1)
# get Hopsworks Model Serving handle
ms = project.get_model_serving()
my_predictor = ms.create_predictor(my_model)
my_deployment = my_predictor.deploy()
print(my_deployment.get_state())
| RETURNS | DESCRIPTION |
|---|---|
|
|