Workshop: Build an AI/ML pipeline with BERT, TensorFlow and Amazon SageMaker
**Description**
In this hands-on workshop, we will build an end-to-end AI/ML pipeline for natural language processing with Amazon SageMaker.
You will learn how to:
• Ingest data into S3 using Amazon Athena and the Parquet data format
• Visualize data with pandas, matplotlib in Jupyter notebooks
• Run data bias analysis with SageMaker Clarify
• Perform feature engineering on a raw dataset using Scikit-Learn and SageMaker Processing Jobs
• Store and share features using SageMaker Feature Store
• Train and evaluate a custom BERT model using TensorFlow, Keras, and SageMaker Training Jobs
• Evaluate the model using SageMaker Processing Jobs
• Track model artifacts using Amazon SageMaker ML Lineage Tracking
• Run model bias and explainability analysis with SageMaker Clarify
• Register and version models using SageMaker Model Registry
• Deploy a model to a REST Inference Endpoint using SageMaker Endpoints
• Automate ML workflow steps by building end-to-end model pipelines using SageMaker Pipelines
**Pre-requisites**
Modern browser – and that’s it!
Every attendee will receive a cloud instance
Nothing will be installed on your local laptop
Everything can be downloaded at the end of the workshop
**Location**
Online
Related Links
=============
O’Reilly Book: https://www.amazon.com/dp/1492079391/
Website: https://datascienceonaws.com
Meetup: https://meetup.datascienceonaws.com
GitHub Repo: https://github.com/data-science-on-aws/
YouTube: https://youtube.datascienceonaws.com
Slideshare: https://slideshare.datascienceonaws.com
Support: https://support.pipeline.ai