Skip to content

ramonamezquita/mlflow-executor

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

End-to-end training and deployment of MLflow projects.

Example

Train and deploy this MLflow example project using the Ray executor.

from mlflow_executor import project, backends, deployers

# Project uri.
uri = "https://github.com/mlflow/mlflow-example"

# Project params.
params = {"alpha": 0.5, "l1_ratio": 0.01}
    
# Project env manager.
# Different options are
#   - local: use the local environment
#   - conda: use conda (must have conda installed).
env_manager = "virtualenv"


# The object ``promise`` contains a reference to the executed task.
promise = project.run(
    uri=uri, 
    parameters=params, 
    env_manager=env_manager 
    backend=backends.RayBackend()
)


# Return the result of the call that ``promise`` represents.
result = promise.result()


# Choose your preferred deployer (e.g., :class:`LocalDeployer`).
# All deployers return a :class:`Predictor` instance that handles
# POST requests to the model inference service.
deployer = deployers.LocalDeployer()
predictor = deployer.deploy(result['model_uri'])

# Check health of the model inference service.
predictor.health()

About

End-to-end training and deployment of MLflow projects.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages