Skip to content

Deployment of XGBoost model using triton server (FIL backend) and Ray Serve

Notifications You must be signed in to change notification settings

shidqiet/xgboost-triton-ray

Repository files navigation

Deployment of XGBoost model using Triton Inference Server and Ray Serve

  • Deploy XGBoost model using Triton Inference Server FIL backend
  • Use Ray Serve to:
    • Handle inference request (HTTP)
    • Preprocess request
    • Pass on the processed request using tritonclient
    • Postprocess and returning inference result

To use this repository, you can follow these steps:

  1. Train and generate model repository directory
python3 train_model.py

In this repository I am not too focused on the model development process, so i just try to train xgboost model on iris dataset.

  1. Build client docker image
./build.sh
  1. Launch triton server and client service
docker-compose up
  1. Test inference
python3 test_infer.py

About

Deployment of XGBoost model using triton server (FIL backend) and Ray Serve

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published