AI Inference Workloads: Overcoming Challenges of taking AI to Production
Inference Workloads: Overcoming Challenges of taking AI to Production
Featured Speaker:  Guy Salton, Global Head of SE, Run:ai
 
Tuesday, May 24, 2022
2:00pm ET; 11:00am PT
60 minutes
Online
AI and MLOps engineers in federal agencies often struggle to deploy models on GPUs. Most Al research initiatives never make it to production. Why? Researchers are facing bottlenecks due to static allocations of GPUs, and different technology sets complicate moving models from training to production.


Join Run:AI and Carahsoft to learn how your agency can overcome the challenges associated with new hardware-accelerated AI modeling practices and discover how traditional best practices have evolved to become more efficient


During this live session, you will learn from our experts how to:

  • Run multiple inference workloads on the same GPU by using the concept of fractional GPUs
  • Remove the bottlenecks which prevent almost 80% of workflows from reaching production
  • Get dynamic MIG slices for each new job when using the NVIDIA A100 GPU
  • Improve GPU utilization when running inference workloads
  • Maintain high throughput and low latency for model serving

Register Now: Learn how you can overcome obstacles involving hardware-accelerated AI modeling practices! 

Unable to attend? 
Sign up here and we'll send you a copy of the recording

carahsoft
Hunter A. Davis
571-662-3927

By supplying my contact information, I authorize Carahsoft and its vendors and partner community to contact me with personalized communications about their products and services. Please review our Privacy Policy for more details or to opt-out at any time.
 
 
 
© 2022 Carahsoft Technology Corp.
11493 Sunset Hills Road
Reston, Virginia  20190
www.carahsoft.com
T: 703.871.8500  |  F: 703.871.8505