The AI chip giant says the open-source software library, TensorRT-LLM, will double the H100’s performance for running inference on leading large language models when it comes out next month.
The H200 features 141GB of HBM3e and a 4.8 TB/s memory bandwidth, a substantial step up from Nvidia’s flagship H100 data center GPU. ‘The integration of faster and more extensive memory will ...
In a statement today, YTL said it will deploy Nvidia H100 Tensor Core GPUs, which power today’s most advanced AI data centres, and use Nvidia AI Enterprise software to streamline production AI.