Corey Nolet’s Post

View profile for Corey Nolet, graphic

Principal Engineer | Big-Data Science, ML and Graph Analytics | High-Performance & Distributed Computing

Device-initiated reads from NVMe that don’t need to wait on CPU scheduling? Yes, please! This is massively useful tech, especially for access patterns that are already optimized for fast disk IO, and when compute can already be moved to the GPU.

View profile for Ryan Meredith, graphic

Director, Storage Solutions Architecture at Micron Technology

NVIDIA GTC'24: Check out Micron's new Gen5 NVMe SSD get 2x the performance on BaM and the NVIDIA H100 get 5x faster GNN training time! Check out our blog here: https://lnkd.in/gms3DjuB Micron Technology worked with teams at Dell Technologies and NVIDIA to produce industry-leading research on AI training model offload to NVMe, which it showcased at the NVIDIA GTC global AI conference. We tested Big Accelerator Memory (BaM) with GPU-initiated direct storage (GIDS) on the NVIDIA H100 Tensor Core GPU in a Dell PowerEdge R7625 server with Micron’s upcoming high-performance Gen5 E3.S NVMe SSD. Huge thanks to the following folks at Micron, Dell, and NVIDIA for making this possible: Micron: John Mazzie, Jeff Armstrong Dell: Seamus Jones, Jeremy Johnson, Mohan Rokkam NVIDIA: Vikram Sharma Mailthody, Chris (CJ) Newburn, Brian Park, Zaid Qureshi, Wen-mei Hwu #MicronTechnology #DellTechnologies #NVIDIA #AI #NVMe #GPU #storage #research #innovation #GTC2024

  • No alternative text description for this image

To view or add a comment, sign in

Explore topics