Fine tune models like LLaMA 2
Optimize Transformers Models and LLMs through efficient processes, and accelerate the training of larger models with the cutting-edge Tensor Cores 4th generation technology and the latest 8-bit data format.
Accelerate your model training and inference with the most high-end AI chip of the market!
Optimize Transformers Models and LLMs through efficient processes, and accelerate the training of larger models with the cutting-edge Tensor Cores 4th generation technology and the latest 8-bit data format.
Accelerate your model serving workloads thanks to Transformer Engine 30x faster for AI inference and new data formats.
With 2nd generation of Secure MIG (multi-instance GPU), partition the GPU into isolated, right-size instances to maximize utilization for the smallest to biggest multi-GPU jobs.
GPUNVIDIA H100 PCIe Tensor Core
GPU Memory80GB HBM2e
Processor24 vCPUs AMD Epyc Zen 4
Processor frequency2.7 Ghz
Memory240 GB of RAM
Memory typeDDR5
Bandwidth10 Gbps
StorageBlock Storage for the boot and 3TB of Scratch Storage NVMe
Understands, interprets, and generates human language in a way that is both meaningful and contextually relevant.
Thanks to models and algorithms specialized in:
Converts spoken language into written text, facilitating the translation of verbal communication into machine-readable data.
Thanks to models and algorithms specialized in:
Generates new content, such as images, text, audio, code. It autonomously produces novel and coherent outputs, expanding the realm of AI-generated content beyond replication or prediction.
With models and algorithms specialized in:
Enables machines to interpret and understand visual information from the world, much like human vision.
Thanks to models and algorithms specialized in:
Predicts and suggests items of interest to users based on their preferences and behaviors, enhancing personalized recommendations and decision-making in various applications.
Using for example Deep Learning Recommendation Model v2 (DLRMv2) that employs DCNv2 cross-layer and a multi-hot dataset synthesized from the Criteo dataset.
Benefit from a ready-to-use Ubuntu image to launch your favorite deep learning containers (pre-installed NVIDIA driver and Docker environment).
Easily launch your favorite JupyterLab or Notebook thanks to the pre-installed Docker environment
Access multiple container registries: your own build containers, Scaleway AI containers, NVIDIA NGC registry and any other registry
Access hundreds of AI softwares optimized by Nvidia to maximise the efficiency of your GPUs and boost your productivity. Among hundreds of softwares developed by NVIDIA and tested by leaders of their industry, harness the efficiency of