Skip to content

Fast tryout scrips for deploying your models with opensource inference engines

Notifications You must be signed in to change notification settings

zymale/model_deployment

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Model Deployment

Scrips for trying out or deploying your models with different deeplearning inference engines.

Typical Model Deployment Pipeline

Folks in Lab often evaluate model in extract the same environment they trained it which is fast and requires little effort but not appropriate for efficient production deployment.

A typical model deployment pipeline in production scale usually involves steps bellows:

  1. Train your model with your favorite deeplearning frameworks (tensorflow/pytorch/caffe)
  2. Export your model to a frozen graph, which can be .pb for tensorflow or .onnx for pytorch.
  3. Convert the frozen graph from step2

Supported inference engines:

Usage

  1. freeze your model and convert it to specific IR
# tf_graph_tookit.py/converter.py provide helpful functions for exporting your model and convert it to IR
  1. Just try out your desired Inference engine
model_dir = '' # your converted IR
ie = SomeIE(model_dir, *args, **kwars)
input_data = None
result = ie.predict(None)

About

Fast tryout scrips for deploying your models with opensource inference engines

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages