Deploying Your Model
TorchServe allows you to expose a WEB API for your Pytorch model that may be accessed directly or via your application.
3 Steps
- Choose a default handler or author a custom model handler.
You will define a class including preprocessing/inference/postprocessing functions. - Create a “model archive(all the files you need for the model to run inference)” and place it in the model store.
- Start TorchServe and load a model archive for serving via the inference API
- Configure and control TorchServe through the Management API.
If you are interested, please check out the Demo too.