Vultr Cloud GPU Accelerated by NVIDIA H200 Tensor Core
Vultr Cloud GPU Accelerated by NVIDIA H200 Tensor Core
The first GPU with HBM3e memory technology delivers 141 GB capacity and 4.8 TB/s bandwidth for AI workloads.
The first GPU with HBM3e memory technology delivers 141 GB capacity and 4.8 TB/s bandwidth for AI workloads.
As AI models grow more sophisticated, enterprises need to minimize training and inference times while keeping infrastructure costs efficient.
The NVIDIA H200 on Vultr features 141 GB of HBM3e memory with 4.8 TB/s bandwidth, delivering up to 2x the LLM inference performance over H100 and 5x faster training versus A100 across 32 global cloud data center regions.

Superior performance for GenAI, LLMs, and high-performance computing workloads
Superior performance for GenAI, LLMs, and high-performance computing workloads
The NVIDIA H200 nearly doubles memory capacity and provides 1.4x the bandwidth of the NVIDIA H100, enabling superior scalability for large-scale AI models, real-time analytics, and scientific simulations while reducing energy consumption and total cost of ownership.
Get started with the world's
largest privately-held cloud
infrastructure company
Create an accountGet started with the world's
largest privately-held cloud
infrastructure company