Temporibus Autem Quibusdam et Aut Debitis Necessitatibus Nam

At vero eos et accusamus et iusto odio dignissimos ducimus qui blanditiis praesentium voluptatum deleniti atque corrupti quos dolores et quas molestias excepturi sint occaecati cupiditate non provident, Similique Sunt in Culpa Qui Officia Deserunt Mollitia Cnimi, id est laborum et dolorum fuga. Et Harum Quidem Rerum Facilis Est Et expedita distinctio Nam libero tempore, cum soluta nobis.

Et harum quidem rerum facilis est et expedita distinctio. Nam libero tempore, cum soluta nobis est eligendi optio cumque nihil impedit quo minus id quod maxime placeat facere possimus, omnis voluptas assumenda est, omnis dolor repellendus. Temporibus autem quibusdam et aut officiis debitis aut rerum.

Nemo enim ipsam voluptatem quia voluptas sit aspernatur aut odit aut fugit, sed quia consequuntur magni dolores eos qui ratione voluptatem sequi nesciunt. Neque porro quisquam est, qui dolorem ipsum quia dolor sit amet, consectetur, adipisci velit, sed quia non numquam eius modi tempora incidunt ut labore et dolore magnam aliquam quaerat voluptatem. Ut enim ad minima veniam, quis nostrum exercitationem ullam. corporis.suscipit laboriosam, nisi ut aliquid ex ea commodi consequatur?

Excepteur sint est non

Quis Nostrud laboris Exercitation ullamco

Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea. commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.

At vero eos et accusamus et iusto odio dignissimos ducimus qui blanditiis praesentium voluptatum deleniti atque corrupti quos dolores et quas molestias excepturi sint occaecati cupiditate non provident, similique sunt. culpa qui officia deserunt mollitia animi, id est laborum et dolorum fuga. Et harum quidem rerum facilis est et expedita distinctio. Nam libero tempore, cum soluta nobis est eligendi optio cumque nihil impedit quo minus id quod maxime placeat facere possimus, omnis voluptas assumenda est, omnis dolor repellendus. Temporibus autem quibusdam et aut officiis debitis aut rerum necessitatibus saepe eveniet ut et voluptates repudiandae sint et molestiae non recusandae. Itaque earum rerum hic tenetur a sapiente delectus, ut aut reiciendis voluptatibus maiores alias consequatur aut perferendis doloribus asperiores repellat

  • At vero eos et accusamus et iusto odio dignissimos
  • Ducimus qui blanditiis praesentium voluptatum Deleniti
  • atque corrupti quos dolores et quas Molestias excepturi
  • sint occaecati cupiditate non Provident, similique
  • sunt culpa qui officia deserunt Mollitia animi, id est laborum et dolorum fuga
100,000+ databases migrated to AWS 10,000+ data lakes and data warehouses deployed on AWS Hundreds of thousands of customers rely on AWS Databases
Build an End-To-End Pipeline With BERT, Tensorflow, and Amazon SageMaker

April 15th, 2021 | 9:00 AM - 1:00 PM PT


In this hands-on workshop, we will build an end-to-end AI/ML pipeline for natural language processing with Amazon SageMaker.


Attendees will learn how to:
  • Ingest data into S3 using Amazon Athena and the Parquet data format
  • Visualize data with pandas, matplotlib on SageMaker notebooks
  • Run data bias analysis with SageMaker Clarify
  • Perform feature engineering on a raw dataset using Scikit-Learn and SageMaker Processing Jobs
  • Store and share features using SageMaker Feature Store
  • Train and evaluate a custom BERT model using TensorFlow, Keras, and SageMaker Training Jobs
  • Evaluate the model using SageMaker Processing Jobs
  • Track model artifacts using Amazon SageMaker ML Lineage Tracking
  • Run model bias and explainability analysis with SageMaker Clarify
  • Register and version models using SageMaker Model Registry
  • Deploy a model to a REST Inference Endpoint using SageMaker Endpoints
  • Automate ML workflow steps by building end-to-end model pipelines using SageMaker Pipelines
If you have any questions about the registration process, please email [email protected].


Event Agenda

  • Workshop introduction and setup.
  • Register and explore our dataset using AWS Glue, Amazon Athena and Parquet data.
  • Perform data bias analysis using SageMaker Clarify.
  • Understand BERT embeddings, and how to convert raw text into BERT features using Hugging Face and TensorFlow. Store and share features using SageMaker feature store.
  • Understand BERT pre-training vs. fine-tuning, and how to fine-tune a BERT model and create a text classifier using Hugging Face and TensorFlow.
  • Build an end-to-end BERT text classifier pipeline.
  • Perform model bias analysis and explainability using SageMaker Clarify.
  • Register and approve model using SageMaker Model Registry, deploy model on REST inference endpoint.

Session Details

  • Track1 Session 1
  • Track2 Session 1
  • Track3 Session 1
  • Track4 Session 1
  • Track5 Session 1
  • Track6 Session 1
  • Track7 Session 1
  • Track8 Session 1
  • Track9 Session 1
  • Track10 Session 1

Track1 Session 1

Migrate your Oracle and SQL Server databases to

Organizations today are looking to free themselves from the constraints of on-premises databases and leverage the power of fully managed databases in the cloud. Amazon RDS is a fully managed relational database service that you can use to run your choice of database engines including open source engines, Oracle, and SQL Server in the cloud. Amazon RDS automates time-consuming database administration tasks and adds capabilities such as replication and Multi-AZ failover to make your database deployments more scalable, available, reliable, manageable, and cost-effective. This session covers why you should consider moving your on-premises Oracle & SQL Server deployments to Amazon RDS and the tools to get started.

Track2 Session 1

Migrate your Oracle and SQL Server databases to

Organizations today are looking to free themselves from the constraints of on-premises databases and leverage the power of fully managed databases in the cloud. Amazon RDS is a fully managed relational database service that you can use to run your choice of database engines including open source engines, Oracle, and SQL Server in the cloud. Amazon RDS automates time-consuming database administration tasks and adds capabilities such as replication and Multi-AZ failover to make your database deployments more scalable, available, reliable, manageable, and cost-effective. This session covers why you should consider moving your on-premises Oracle & SQL Server deployments to Amazon RDS and the tools to get started.

Track3 Session 1

Tab 3 content

Nulla eleifend felis vitae velit tristique imperdiet. Etiam nec imperdiet elit. Pellentesque sem lorem, scelerisque sed facilisis sed, vestibulum sit amet eros.

Track4 Session 1

Tab 4 content

Integer ultrices lacus sit amet lorem viverra consequat. Vivamus lacinia interdum sapien non faucibus. Maecenas bibendum, lectus at ultrices viverra, elit magna egestas magna, a adipiscing mauris justo nec eros.

Track5 Session 1

Tab 5 content

Nulla eleifend felis vitae velit tristique imperdiet. Etiam nec imperdiet elit. Pellentesque sem lorem, scelerisque sed facilisis sed, vestibulum sit amet eros.

Track6 Session 1

Tab 6 content

Integer ultrices lacus sit amet lorem viverra consequat. Vivamus lacinia interdum sapien non faucibus. Maecenas bibendum, lectus at ultrices viverra, elit magna egestas magna, a adipiscing mauris justo nec eros.

Track7 Session 1

Tab 7 content

Organizations today are looking to free themselves from the constraints of on-premises databases and leverage the power of fully managed databases in the cloud. Amazon RDS is a fully managed relational database service that you can use to run your choice of database engines including open source engines, Oracle, and SQL Server in the cloud. Amazon RDS automates time-consuming database administration tasks and adds capabilities such as replication and Multi-AZ failover to make your database deployments more scalable, available, reliable, manageable, and cost-effective. This session covers why you should consider moving your on-premises Oracle & SQL Server deployments to Amazon RDS and the tools to get started.

Track8 Session 1

Tab 8 content

Organizations today are looking to free themselves from the constraints of on-premises databases and leverage the power of fully managed databases in the cloud. Amazon RDS is a fully managed relational database service that you can use to run your choice of database engines including open source engines, Oracle, and SQL Server in the cloud. Amazon RDS automates time-consuming database administration tasks and adds capabilities such as replication and Multi-AZ failover to make your database deployments more scalable, available, reliable, manageable, and cost-effective. This session covers why you should consider moving your on-premises Oracle & SQL Server deployments to Amazon RDS and the tools to get started.

Track9 Session 1

Tab 9 content

Organizations today are looking to free themselves from the constraints of on-premises databases and leverage the power of fully managed databases in the cloud. Amazon RDS is a fully managed relational database service that you can use to run your choice of database engines including open source engines, Oracle, and SQL Server in the cloud. Amazon RDS automates time-consuming database administration tasks and adds capabilities such as replication and Multi-AZ failover to make your database deployments more scalable, available, reliable, manageable, and cost-effective. This session covers why you should consider moving your on-premises Oracle & SQL Server deployments to Amazon RDS and the tools to get started.

Track10 Session 1

Tab 10 content

Organizations today are looking to free themselves from the constraints of on-premises databases and leverage the power of fully managed databases in the cloud. Amazon RDS is a fully managed relational database service that you can use to run your choice of database engines including open source engines, Oracle, and SQL Server in the cloud. Amazon RDS automates time-consuming database administration tasks and adds capabilities such as replication and Multi-AZ failover to make your database deployments more scalable, available, reliable, manageable, and cost-effective. This session covers why you should consider moving your on-premises Oracle & SQL Server deployments to Amazon RDS and the tools to get started.

Multiple City Registration

  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa
  • JAN 21 2020
    Sunt in Culpa

Venue

The Meteropolitian
233 south wacker drive, 67th floor, Chicago, Illinois 60606

Sessions will focus on providing an overview of AWS services and features, with the assumption that attendees are new to the topic Sessions will focus on providing an overview of AWS services and features, with the assumption that attendees are new to the topic Sessions will focus on providing an overview of AWS services and features, with the assumption that attendees are new to the topic Sessions will focus on providing an overview of AWS services and features, with the assumption that attendees are new to the topic

Session Proficiency Levels Explained

  • Level 100

    Introductory

    Sessions will focus on providing an overview of AWS services and features, with the assumption that attendees are new to the topic

  • Level 200

    Intermediate

    Sessions will focus on providing best practices, details of service features and demos with the assumption that attendees have introductory knowledge of the topics

  • Level 300

    Advanced

    Sessions will dive deeper into the selected topic. Presenters assume that the audience has some familiarity with the topic, but may or may not have direct experience implementing a similar solution

  • Level 400

    Expert

    Sessions are for attendees who are deeply familiar with the topic, have implemented a solution on their own already, and are comfortable with how the technology works across multiple services, architectures, and implementations

Why should you attend?

This event provides attendees with the chance to deeply explore some AWS AI/ML service offerings through in-depth demos and advanced technically focused sessions.

Who should attend Dev Days?

Anyone who would like to learn from a technical expert deep dive in AWS Services. Technical individuals that manipulate code, data, configuration/operations or design app architectures to build differentiated and successful applications in the Cloud, such as developers, data engineers, data scientists, enterprise application architects, IT admins/operators, infosec, or those experimenting with emerging tech.

Featured Speakers

  • EV_cfregly_200x200_Apr-2021.png

    Chris Fregly,
    Principal Developer Advocate, AI and Machine Learning, AWS

    Chris is a Developer Advocate for AI and Machine Learning at Amazon Web Services (AWS) based in San Francisco, California. He is co-author of the O'Reilly Book, "Data Science on AWS". Chris is also the founder of many AI-focused global meetups including the global "Data Science on AWS" Meetup. He regularly speaks at AI and Machine Learning conferences across the world including O’Reilly AI, Open Data Science Conference (ODSC), and Nvidia GPU Technology Conference (GTC).

    Previously, Chris was Founder at PipelineAI where he worked with many AI-first startups and enterprises to continuously deploy ML/AI Pipelines using Spark ML, Kubernetes, TensorFlow, Kubeflow, Amazon EKS, and Amazon SageMaker.

  • EV_abarth_200x200_Apr-2021.jpg

    Antje Barth,
    Senior Developer Advocate, AI and Machine Learning, AWS

    Antje is a Developer Advocate for AI and Machine Learning at Amazon Web Services (AWS) based in Düsseldorf, Germany. She is co-author of the O'Reilly Book, "Data Science on AWS".

    Antje is also co-founder of the Düsseldorf chapter of Women in Big Data. She frequently speaks at AI and Machine Learning conferences and meetups around the world, including the O’Reilly AI and Strata conferences. Besides ML/AI, Antje is passionate about helping developers leverage Big Data, container and Kubernetes platforms in the context of AI and Machine Learning.

    Previously, Antje worked in technical evangelism and solutions engineering at MapR and Cisco where she worked with many companies to build and deploy cloud-based AI solutions using AWS and Kubernetes.

  • Richard Boyd, Cloud Data Engineer, iRobot

    Richard Boyd is a cloud data engineer with the iRobot Corporation’s Cloud Data Platform where he builds tools and services to support the world’s most beloved vacuum cleaner. Before joining iRobot, Richard built discrete event simulators for Amazon’s automated fulfillment centers in Amazon Robotics. His previous roles include cyber warfare systems analyst at MIT and research for the Center for Army Analysis. He holds advanced degrees in Applied Mathematics & Statistics.

  • Raju Gulabani, VP of Databases, Analytics & AI, AWS

    Raju Gulabani is VP of Databases, Analytics & AI within AWS at Amazon.com. He is responsible for P&<, product management, engineering and operations for Database services such as Amazon Aurora and Amazon DynamoDB, and Analytics services such as Amazon Redshift and Amazon EMR, as well as AI services like Amazon Lex, Amazon Polly, and Amazon Rekognition. Prior to joining Amazon in his current position in 2010, Raju spent four years at Google and built the Google Apps business (now known as G Suite).Earlier in his career, Raju founded an Intel backed Wi-Fi Voice over IP company as well as held engineering management positions at Microsoft.

  • Ryan Kelly, Data Architect, Equinox

    Ryan Kelly is a data architect at Equinox, where he helps outline and implement frameworks for data initiatives. He also leads clickstream tracking which helps aid teams with insights on their digital initiatives. Ryan loves making it easier for people to reach and ingest their data for the purposes of business intelligence, analytics, and product/service enrichment. He also loves exploring and vetting new technologies to see how they can enhance what they do at Equinox

  • Richard Boyd, Cloud Data Engineer, iRobot

    Richard Boyd is a cloud data engineer with the iRobot Corporation’s Cloud Data Platform where he builds tools and services to support the world’s most beloved vacuum cleaner. Before joining iRobot, Richard built discrete event simulators for Amazon’s automated fulfillment centers in Amazon Robotics. His previous roles include cyber warfare systems analyst at MIT and research for the Center for Army Analysis. He holds advanced degrees in Applied Mathematics & Statistics.

  • Raju Gulabani, VP of Databases, Analytics & AI, AWS

    Raju Gulabani is VP of Databases, Analytics & AI within AWS at Amazon.com. He is responsible for P&<, product management, engineering and operations for Database services such as Amazon Aurora and Amazon DynamoDB, and Analytics services such as Amazon Redshift and Amazon EMR, as well as AI services like Amazon Lex, Amazon Polly, and Amazon Rekognition. Prior to joining Amazon in his current position in 2010, Raju spent four years at Google and built the Google Apps business (now known as G Suite).Earlier in his career, Raju founded an Intel backed Wi-Fi Voice over IP company as well as held engineering management positions at Microsoft.

  • Ryan Kelly, Data Architect, Equinox

    Ryan Kelly is a data architect at Equinox, where he helps outline and implement frameworks for data initiatives. He also leads clickstream tracking which helps aid teams with insights on their digital initiatives. Ryan loves making it easier for people to reach and ingest their data for the purposes of business intelligence, analytics, and product/service enrichment. He also loves exploring and vetting new technologies to see how they can enhance what they do at Equinox

  • Richard Boyd, Cloud Data Engineer, iRobot

    Richard Boyd is a cloud data engineer with the iRobot Corporation’s Cloud Data Platform where he builds tools and services to support the world’s most beloved vacuum cleaner. Before joining iRobot, Richard built discrete event simulators for Amazon’s automated fulfillment centers in Amazon Robotics. His previous roles include cyber warfare systems analyst at MIT and research for the Center for Army Analysis. He holds advanced degrees in Applied Mathematics & Statistics.

  • Raju Gulabani, VP of Databases, Analytics & AI, AWS

    Raju Gulabani is VP of Databases, Analytics & AI within AWS at Amazon.com. He is responsible for P&<, product management, engineering and operations for Database services such as Amazon Aurora and Amazon DynamoDB, and Analytics services such as Amazon Redshift and Amazon EMR, as well as AI services like Amazon Lex, Amazon Polly, and Amazon Rekognition. Prior to joining Amazon in his current position in 2010, Raju spent four years at Google and built the Google Apps business (now known as G Suite).Earlier in his career, Raju founded an Intel backed Wi-Fi Voice over IP company as well as held engineering management positions at Microsoft.

  • Ryan Kelly, Data Architect, Equinox

    Ryan Kelly is a data architect at Equinox, where he helps outline and implement frameworks for data initiatives. He also leads clickstream tracking which helps aid teams with insights on their digital initiatives. Ryan loves making it easier for people to reach and ingest their data for the purposes of business intelligence, analytics, and product/service enrichment. He also loves exploring and vetting new technologies to see how they can enhance what they do at Equinox

  • Richard Boyd, Cloud Data Engineer, iRobot

    Richard Boyd is a cloud data engineer with the iRobot Corporation’s Cloud Data Platform where he builds tools and services to support the world’s most beloved vacuum cleaner. Before joining iRobot, Richard built discrete event simulators for Amazon’s automated fulfillment centers in Amazon Robotics. His previous roles include cyber warfare systems analyst at MIT and research for the Center for Army Analysis. He holds advanced degrees in Applied Mathematics & Statistics.

Customer Highlights

Epic Logo Image

Epics Games’ entire analytics platform runs on AWS. Billions of game events, like player interactions on the map, their accuracy, damage taken and dealt, and what resources they are using are all sent to AWS.

Epics Games

Yelp Logo Image

Epics Games’ entire analytics platform runs on AWS. Billions of game events, like player interactions on the map, their accuracy, damage taken and dealt, and what resources they are using are all sent to AWS.

Yelp

Airbnb Logo Image

Epics Games’ entire analytics platform runs on AWS. Billions of game events, like player interactions on the map, their accuracy, damage taken and dealt, and what resources they are using are all sent to AWS.

Airbnb

Lyft Image

Epics Games’ entire analytics platform runs on AWS. Billions of game events, like player interactions on the map, their accuracy, damage taken and dealt, and what resources they are using are all sent to AWS.

Lyft

FAQS

  1. Where is AWS Innovate hosted?
  2. What is the price of attending AWS Innovate?
  3. Who should attend AWS Innovate?
  4. How do I get the certificate of attendance?
Q: Where is this event?

A: This event is an online event, hosted by AWS on the INXPO platform.

Q: Who should attend this event?

A: Developers building data-driven apps; DBAs and data engineers who are building analytics infrastructure and data pipelines; Analysts and data scientists who are deriving insights that answer complex business quesions and building/trainining machine learning models.

Q: How much does this event cost?

A: There is no cost to attend this event.

Q: What are the prerequisites before attending the event?

A: There are no prerequisites for attending the event. We encourage attendees to browse the Database and Analytics pages on the AWS website to get a brief overview of the services available to them.