Nebius monthly digest, August 2024

We recently reduced the prices of NVIDIA H100, invited everyone interested to reserve H200, launched a public preview of Managed Service for MLflow and released a new talk about deploying generative AI models in production.

Price drop: Get NVIDIA H100 GPUs for less

One VM equipped with H100 is now priced at $3.50 per hour when paying as you go, down from the previous price of $4.85.

Therefore, the prices for VMs and clusters reserved for 3, 6 or 12 months are also down. For instance, you can get a cluster for 12 months at less than 3 dollars per H100 GPU hour.

Check out the new pricing and get your cluster right away.

This fall, we’re adding H200 SXM GPUs to Nebius

Currently, H200 is the most powerful GPU for your AI and HPC. Its key gains come from:

  • Memory. H200 is the first GPU with 141 GB of HBM3e memory — nearly doubling the memory capacity of H100 SXM.
  • Performance. The increased memory bandwidth of 4.8 TB/s allows for better utilization of processing power, making it the new preferred GPU for large models.
  • Data access speed. With the shared filesystem, you can achieve up to 20 GB/s read speeds from one node, which is crucial for training and inference. This is 6 times more than what’s available with the H100.

According to an NVIDIA research, H200 shows up to 2x the LLM inference performance over H100: 1.4x in case of Llama2 13B, 1.6x with GPT-3 175B, and 1.9x with Llama2 70B.

Prices start from $2.50 per GPU hour. You can reserve these new powerhouse GPUs today.

Managed Service for MLflow is now in public preview

MLflow is a renowned industry tool that streamlines workflows in the model development cycle. We made MLflow more accessible to a broad audience of ML enthusiasts by providing it as a managed solution.

You can learn more or request access here.

Talk by Nebius: what it takes to train a GenAI model

Several weeks ago, our Product Director Narek Tatevosyan gave a talk at The AI Summit Series in London. He shared an insider’s perspective on the key steps, essential tools and challenges involved in bringing a generative AI model from concept to production. In this talk, he broke down the real story behind data preparation, experiments to prepare foundational model training, a pre-training process, fine-tuning and inference.

The latest on our blog

author
Nebius team
Sign in to save this post