I. Introduction I. Overview of the tutorial II. Importance of serving machine learning models in production III. Brief introduction to Jupyter Notebook and its limitations in production environments
II. Introduction to FastAPI I. What is FastAPI? II. Key features and advantages III. Comparison with other web frameworks IV. Demonstration of basic FastAPI application setup.
III. Serving Machine Learning Models with FastAPI I. Why FastAPI for serving ML models? II. Design considerations for integrating ML models with FastAPI III. Demonstration of serving a machine learning model as a REST API with FastAPI IV. Handling model inputs and outputs
IV. Containerization with Docker I. Introduction to Docker II. Benefits of containerization for deploying ML models III. Dockerizing a FastAPI application IV. Best practices for Dockerizing machine learning applications
V. Deployment with Google Cloud Run I. Introduction to Google Cloud Run II. Advantages of using Cloud Run for deploying containerized applications III. Step-by-step guide to deploying a Dockerized FastAPI application on Google Cloud Run IV. Configuring scalability and monitoring for deployed applications
VI. Best Practice: Security, Performance Optimisation I. Security considerations for deploying ML models in production II. Performance optimization techniques III. Handling versioning and updates of deployed models
VII. Conclusion I. Recap of key takeaways II. Encouragement for attendees to explore serving ML models beyond Jupyter Notebook III. Resources for further learning and exploration IV. Questions and answers