Bare Metal GPUs: Dedicated machines for advanced AI workloads

Seriously powerful compute, purpose-built for your most demanding AI/ML projects.

Coming Soon: Bare Metal H200 GPUs

Powered by NVIDIA H200 Tensor Core GPUs with HBM3e, Bare Metal servers deliver exceptional performance and memory for generative AI, large language models, and high-performance computing (HPC) workloads.

Contact our team to learn more

How our customers use Bare Metal GPUs

Reserve capacity now

Lepton AI

From the creators of Caffe, Pytorch, ONNX, and etcd.

Lepton AI provides a cutting-edge AI inference and training platform. They focus on optimizing the deployment and utilization of GPUs in various compute environments. Efficiency, reliability and ease of use are at the core of their software design.

Supermaven

From the creator of Tabnine

Supermaven's AI code completion tool boosts coding speed and efficiency with high-quality suggestions, in-editor AI chat, and a large context window of up to 1 million tokens. Powered by their advanced LLM, Babble, it's 2.5x larger than their previous model.

Moonvalley

Moonvalley AI specializes in generative media, creating cinematic videos and animations from text and image prompts. Their technology enables fast, high-definition, 16:9 video production, reducing typical production time and effort. With over 300,000 users, they operate through a Discord server and Easy With AI.

High-Performance Bare Metal GPUs for AI/ML Workloads

For serious AI builders who need more control, power, and less noise in their cloud infrastructure, DigitalOcean Bare Metal GPUs with NVIDIA accelerated computing are ready to deploy. Purpose-built for the most demanding AI/ML workloads and featuring 8 NVIDIA Hopper GPUs and powerful hardware, these servers offer high-performance compute capabilities and customizable options designed for your most intense processing needs.
DigitalOcean Bare Metal GPUs provide full access to all GPUs, offering dedicated, single-tenant infrastructure with no neighbors that makes them ideal for large-scale model training, real-time inference, and complex orchestration. Take direct control over Bare Metal GPU hardware to ensure highest performance and privacy. With Bare Metal GPUs, you can fully optimize model training, fine-tune pre-trained models, achieve high-speed inference, and build scalable custom architectures for specialized use cases.

Power your projects with the flexibility of Bare Metal GPUs

Platform Reliability

Get more reliable, high-performance compute for demanding workloads.

Support

Receive high-touch engineering support, helping to ensure smooth operations and peak server performance.

Pricing

Enjoy cost-efficient pricing that includes storage, and favorably compares to other providers.

Availability

Bare Metal GPUs are available in New York, USA and Amsterdam, Netherlands, with more data centers coming soon.

Trusted by developers for fast performance and full control

Bare Metal GPUs vs. GPU Droplets

Bare Metal GPUs and GPU Droplets are suited to different workloads. GPU Droplets offer easy scalability and fast provisioning for tasks like LLM training, model fine-tuning, or inference. Bare Metal GPUs get you maximum performance and complete control over hardware, ideal for high-throughput workloads.

Reserve capacity now

Bare Metal H100 Servers

GPU Droplet H100

GPU Droplet H100x8

GPU Memory(v)CPUsCPU TypeNetwork CardsMachine MemoryLocal Storage (Boot)Local Storage (Scratch)
640 GB1922x Intel Xeon Platinum 8468Nvidia/Mellanox ConnectX-7 8x400G NIC
Nvidia/Mellanox ConnectX-6 Dx 4x100GE NIC
2 TiB1,024 GiB NVMe (RAID1)35 TiB NVMe (RAID5)
80 GB202x Intel Xeon Platinum 8458PNvidia/Mellanox ConnectX-6 2x100GE NIC240 GiB720 GiB NVMe5 TiB NVMe
640 GB1602x Intel Xeon Platinum 8458PNvidia/Mellanox ConnectX-6 2x100GE NIC1,920 GiB2 TiB NVMe40 TiB NVMe

Bare Metal GPU Resources

What are Bare Metal GPUs?

What are Bare Metal GPUs?

GPU Memory bandwidth

GPU Memory bandwidth

Monitoring GPU utilization for Deep Learning

Monitoring GPU utilization for Deep Learning

Multi-GPU on raw PyTorch with Hugging Face's Accelerate library

Multi-GPU on raw PyTorch with Hugging Face's Accelerate library

How to maximize GPU utilization by finding the right batch size

How to maximize GPU utilization by finding the right batch size

The Hidden Bottleneck: How GPU Memory Hierarchy Affects Your Computing Experience

The Hidden Bottleneck: How GPU Memory Hierarchy Affects Your Computing Experience

Frequently asked questions

What is a Bare Metal GPU?

A Bare Metal GPU is a physical server equipped with powerful GPUs, like the NVIDIA H100. This gives you full control over the hardware so you can install your preferred operating system. Bare Metal GPUs are isolated from other servers, providing excellent security. The server can be used for AI/ML tasks like model training, fine-tuning, and inference.

What are the advantages of Bare Metal GPUs?
  • Full control: Bare Metal provides dedicated, single-tenant hardware, offering full control over the operating system, software, and configurations.
  • High performance: It delivers peak GPU performance, ideal for high-scale AI/ML model training, complex computations, and real-time inference.
  • Customizability: You can deeply customize the hardware and software environment, enabling tailored setups like Kubernetes clusters or other custom orchestration needs.
  • Resource isolation: It offers complete isolation from other users, eliminating the risks of 'noisy neighbors' and helping to ensure security and privacy.
  • Scalability: Bare Metal can be configured for single-node or multi-node setups, allowing for large-scale distributed training and high-performance workloads.
How does Bare Metal work?

Bare Metal works by providing a physical, dedicated server where you can install and run your applications. This setup is particularly useful if you need stable, high-performance infrastructure for demanding workloads like AI/ML model training or custom orchestration.

What advanced use-cases does Bare Metal support?

Bare Metal supports a wide range of use-cases, including:


  • AI/ML workloads: Model training, fine-tuning, and inference, especially for large-scale data processing.
  • Custom orchestrations: Technologies like Kubernetes for containerized environments or other custom setups for complex applications.
  • High-performance computing: Applications that require dedicated resources, such as simulations, scientific computations, and real-time data processing.