
DeepInfra
Best for LLM Serverless APIs, Fast Image Generation, Voice AI

Best for Bare Metal GPU, Low-Latency AI Inference, Global Edge AI Deployment
Latitude.sh is a global bare metal cloud provider founded in 2020, offering dedicated GPU servers across 6 continents. Unlike VM-based GPU clouds, Latitude.sh provides true bare metal servers with no hypervisor overhead, delivering maximum GPU performance and deterministic latency.
| GPU Models | H100 SXM5 80GB, A100 SXM4 80GB, RTX 4090 24GB, A10G 24GB |
| GPU Types | A100, A10G, H100, RTX 4090 |
| Headquarters | San Francisco, CA, USA |
| Founded | 2020 |
| Availability | Available Now |
| Website | latitude.sh ↗ |
💡 Pricing note: Rates shown are indicative. Final pricing depends on GPU model, reservation type (spot vs. on-demand), contract length, and region. Get an exact quote →
Latitude.sh GPU cloud pricing starts from $1.20/hr depending on GPU type, reservation model (on-demand vs. spot vs. reserved), and region. Use the quote form to get exact pricing for your specific workload.
Latitude.sh offers H100 SXM5 80GB, A100 SXM4 80GB, RTX 4090 24GB, A10G 24GB GPU instances. Availability varies by region and configuration. Contact the provider through ComputeStacker for current availability.
Latitude.sh operates data centers in EU West, Singapore, US East, US West. Choosing a region close to your users minimises latency and can help with data residency compliance requirements.
Use the "Get a Quote" button on this page to submit your GPU requirements. ComputeStacker will forward your request to Latitude.sh and other matching providers. You'll receive proposals within 24 hours — no commitment required.
Latitude.sh offers high-performance GPU infrastructure suitable for large language model training and fine-tuning workloads. For large-scale distributed training, check the Specs tab for NVLink and InfiniBand interconnect availability.

Best for LLM Serverless APIs, Fast Image Generation, Voice AI

Best for On-demand GPU instances, SMEs, Sustainable Computing

Best for Serverless Inference, Ad-hoc Python scripts, Quick Prototyping