Broadcast Date: 21 April, 2020

Level: 200

In deep learning applications, inference accounts for up to 90% of compute cost. To reduce this high inference cost, you can use Amazon Elastic Inference, which allows you to attach just the right amount of GPU-powered inference acceleration to any EC2 or SageMaker instance type or ECS task. In this tech talk, you will learn about how to use Elastic Inference for deploying models built on PyTorch, a popular machine learning framework.

Learning Objectives

  • Get an overview of Amazon Elastic Inference
  • Learn about how to use Elastic Inference to reduce costs and improve latency for your PyTorch models on Amazon SageMaker
  • Get a demo using TorchScript with Elastic Inference API

Who Should Attend?

Data Scientists, ML Developers. Researchers, and Data Engineers

Speakers

  • David Thomas, Software Development Engineer, AWS


Learn More

To learn more about the services featured in this talk, please visit:
https://aws.amazon.com/machine-learning/elastic-inference/

Intro body copy here about 2018 re:Invent launches.

Download the Slide Deck

Compute

Service How To

December 19th, 2018 | 1:00 PM PT

Developing Deep Learning Models for Computer Vision with
Amazon EC2 P3 Instances.

Register Now>

Containers

What's New / Cloud Innovation

December 11th, 2018 | 1:00 PM PT

EMBARGOED

Register Now>

Data Lakes & Analytics

Webinar 1:

What's New / Cloud Innovation

December 10th, 2018 | 11:00 AM PT

EMBARGOED

Register Now>

Webinar 2:

What's New / Cloud Innovation

December 12th, 2018 | 11:00 AM PT

EMBARGOED

Register Now>