Deploying AI Models

July 17, 2019

Here’s a discussion that we had about how to deploy models to be used in an AI application.

The thought process follows a pattern of providing a web interface of some sort with the requests being queued up and handled asynchrounously by a server running an AI model.


  • Welcome & Introduction
  • Project Updates
  • Deploying AI Models
  • Plan for next week

Deployment Use Case

Let’s start by thinking about this use case from the Bug Analysis Project

After we train some sort of model, how do we make it available for the masses?

Think about something like this:


How would we build this?

What about updating the model after you deploy it?


Train and serve a TensorFlow model with TensorFlow Serving

fastai serving

There are two very different ways to deploy ML models, here’s both