Accelerating Generative AI

Workloads using generative artificial intelligence trained on large language models are frequently throttled by insufficient resources (e.g., memory, storage, compute or network dataflow bottlenecks). If not identified and addressed, these dataflow bottlenecks can constrain Gen AI application performance well below optimal levels.

Given the compelling uses across natural language processing (NLP), video analytics, document resource development, image processing, image generation, and text generation, being able to run these workloads efficiently has become critical to many IT and industry segments. The resources that contribute to generative AI performance and efficiency include CPUs, DPUs, GPUs, FPGAs, plus memory and storage controllers.

On January 24, 2024, the SNIA Networking Storage Forum (NSF) is convening a panel of experts for a discussion on how to tackle Gen AI challenges at our live webinar, “Accelerating Generative AI – Options for Conquering the Dataflow Bottlenecks,” where a broad cross-section of industry veterans will provide insight into the following:

  • Defining the Gen AI dataflow bottlenecks
  • Tools and methods for identifying acceleration options
  • Matchmaking the right xPU solution to the target Gen AI workload(s)
  • Optimizing the network to support acceleration options
  • Moving data closer to processing, or processing closer to data
  • The role of the software stack in determining Gen AI performance

This is a session you don’t want to miss! Register today to save your spot.

Leave a Reply

Your email address will not be published. Required fields are marked *