Webinar Series Australia and New Zealand

First Name
Last Name
Company Name
CDN Province
US State
India State
AU State
Postal Code
Phone Number
Job Role
This information is associated with my:
Compliance Opt-in
Thank you!
Error - something went wrong!

Scaling LLM/GenAI deployment with NVIDIA Triton on Amazon EKS

Triton is open source inference serving software that simplifies the inference serving process and provides high inference performance. This session explores the synergy of NVIDIA Triton and Amazon EKS for efficient, large-scale machine learning model deployment. We discuss how Triton Inference Server standardizes machine learning inferencing across various deep learning frameworks like PyTorch, ONNX, and TensorRT to streamline GenAI/LLM model deployment.

Level: L400
Speaker: Keita Watanabe, Senior Solutions Architect, Frameworks, AWS

Previous Video
Serverless Developer Experience Workshop
Serverless Developer Experience Workshop

The AWS Serverless Developer Experience workshop provides you with an immersive experience as a serverless ...

Next Video
Next generation Authz with Cedar
Next generation Authz with Cedar

Authorisation is one of the foundational needs when building your applications and services. Making sure yo...