Remove platform nvidia
article thumbnail

AWS at NVIDIA GTC 2024: Accelerate innovation with generative AI on AWS

AWS Machine Learning

AWS was delighted to present to and connect with over 18,000 in-person and 267,000 virtual attendees at NVIDIA GTC, a global artificial intelligence (AI) conference that took place March 2024 in San Jose, California, returning to a hybrid, in-person experience for the first time since 2019.

article thumbnail

Introducing three new NVIDIA GPU-based Amazon EC2 instances

AWS Machine Learning

We are excited to announce the expansion of this portfolio with three new instances featuring the latest NVIDIA GPUs: Amazon EC2 P5e instances powered by NVIDIA H200 GPUs, Amazon EC2 G6 instances featuring NVIDIA L4 GPUs, and Amazon EC2 G6e instances powered by NVIDIA L40S GPUs. times larger and 1.4

Training 106
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

How Amazon Music uses SageMaker with NVIDIA to optimize ML training and inference performance and cost

AWS Machine Learning

Delivering a superior customer experience to instantly find the music that users search for requires a platform that is both smart and responsive. In this post, we walk through the journey Amazon Music took to optimize performance and cost using SageMaker and NVIDIA Triton Inference Server and TensorRT.

article thumbnail

Build a medical imaging AI inference pipeline with MONAI Deploy on AWS

AWS Machine Learning

This post is cowritten with Ming (Melvin) Qin, David Bericat and Brad Genereaux from NVIDIA. AWS and NVIDIA have come together to make this vision a reality. A SageMaker multi-model endpoint uses NVIDIA Triton Inference Server with GPU to run multiple deep learning model inferences.

article thumbnail

A secure approach to generative AI with AWS

AWS Machine Learning

We plan to offer this end-to-end encrypted flow in the upcoming AWS-designed Trainium2 as well as GPU instances based on NVIDIA’s upcoming Blackwell architecture, which both offer secure communications between devices, the third principle of Secure AI Infrastructure.

article thumbnail

How Patsnap used GPT-2 inference on Amazon SageMaker with low latency and cost

AWS Machine Learning

Recently, the AWS Generative AI Innovation Center collaborated with Patsnap to implement a feature to automatically suggest search keywords as an innovation exploration to improve user experiences on their platform. Patsnap provides a global one-stop platform for patent search, analysis, and management. client('sts').get_caller_identity()['Account']

article thumbnail

Serve multiple models with Amazon SageMaker and Triton Inference Server

AWS Machine Learning

In 2021, AWS announced the integration of NVIDIA Triton Inference Server in SageMaker. You can use NVIDIA Triton Inference Server to serve models for inference in SageMaker. In this post, we discuss how SageMaker and NVIDIA Triton Inference Server can solve this problem. nvidia/pytorch:22.04-py3 Solution overview.