- Deploy XGBoost model using Triton Inference Server FIL backend
- Use Ray Serve to:
- Handle inference request (HTTP)
- Preprocess request
- Pass on the processed request using tritonclient
- Postprocess and returning inference result
To use this repository, you can follow these steps:
- Train and generate model repository directory
python3 train_model.py
In this repository I am not too focused on the model development process, so i just try to train xgboost model on iris dataset.
- Build client docker image
./build.sh
- Launch triton server and client service
docker-compose up
- Test inference
python3 test_infer.py