Cloud GPUs for Deep Learning: Availability& Price / Performance
If you are flexible about the GPU model, identify the most cost-effective cloud GPU
If you prefer a specific model (e.g. A100), identify the GPU cloud providers offering it.
If undecided between on-prem and the cloud, explore whether to buy or rent GPUs on the cloud.
Cloud GPU price per throughput
Two common pricing models for GPUs are “on-demand” and “spot” instances.
On-demand GPUs from big tech cloud providers
This is the most straightforward pricing model where you pay for the compute capacity by the hour or second, depending on what you use with no long-term commitments or upfront payments. These instances are recommended for users who prefer the flexibility of a cloud GPU platform without any up-front payment or long-term commitment. On-demand instances are usually more expensive than spot instances, but they provide guaranteed uninterrupted capacity.
Cloud | GPU Type / Memory* | # of GPUs | On-demand $ | Throughput** | Throughput** / $*** |
---|---|---|---|---|---|
Azure | A100 / 80 GB | 1 | 3.67 | 232 | 63 |
Azure | A100 / 80 GB | 4 | 14.69 | 821 | 56 |
Azure | A100 / 80 GB | 2 | 7.35 | 406 | 55 |
GCP | A100 / 40 GB | 1 | 3.67 | 179 | 49 |
Azure | A100 / 80 GB | 8 | 37.18 | 1,362 | 37 |
AWS | A100 / 80 GB | 8 | 40.97 | 1,362 | 33 |
GCP | V100 / 16 GB | 1 | 2.95 | 42 | 14 |
AWS | V100 / 16 GB | 1 | 3.06 | 42 | 14 |
Azure | V100 / 16 GB | 1 | 3.06 | 42 | 14 |
GCP | V100 / 16 GB | 2 | 5.91 | 77 | 13 |
GCP | V100 / 16 GB | 4 | 11.81 | 153 | 13 |
AWS | V100 / 16 GB | 4 | 12.24 | 153 | 13 |
Azure | V100 / 16 GB | 2 | 6.12 | 77 | 13 |
Azure | V100 / 16 GB | 4 | 12.24 | 153 | 13 |
GCP | V100 / 16 GB | 8 | 23.63 | 289 | 12 |
AWS | V100 / 16 GB | 8 | 24.48 | 289 | 12 |
* Memory and GPU model are not the only parameters. CPUs and RAM can also be important, however, they are not the primary criteria that shape cloud GPU performance. Therefore, for simplicity, we have not included number of CPUs or RAM in these tables.
** Training throughput is a good metric to measure relative GPU effectiveness. It measures the number of tokens processed per second by the GPU for a language model (i.e. bert_base_squad).1 Please note that these throughput values should serve as high level guidelines. The same hardware would have a significantly different throughput for your workload since there is significant throughput difference even between LLMs running on the same hardware.2
*** Excludes cost of storage, network performance, ingress/egress etc. This is only the GPU cost.3
On-demand GPUs from other cloud providers
Cloud | GPU Type / Memory* | # of GPUs | On-demand $ | Throughput** | Throughput** / $*** |
---|---|---|---|---|---|
TensorDock | A100 / 80 GB | 4 | 1.200 | 821 | 684 |
Vast.ai | V100 / 16 GB | 8 | 0.944 | 289 | 306 |
Vast.ai | V100 / 16 GB | 2 | 0.358 | 77 | 215 |
TensorDock | A100 / 80 GB | 1 | 1.200 | 232 | 193 |
TensorDock | V100 / 16 GB | 1 | 0.220 | 42 | 191 |
TensorDock | A100 / 80 GB | 1 | 1.400 | 232 | 165 |
Jarvislabs | A100 / 40 GB | 1 | 1.1 | 179 | 163 |
Lambda | A100 / 40 GB | 1 | 1.1 | 179 | 163 |
Lambda | H100 / 80 GB | 1 | 1.99 | 322 | 162 |
Crusoe Cloud | A100 / 80 GB | 1 | 1.650 | 232 | 140 |
FluidStack | A100 / 40 GB | 1 | 1.290 | 179 | 139 |
FluidStack | A100 / 40 GB | 1 | 1.400 | 179 | 128 |
Vast.ai | A100 / 40 GB | 1 | 1.400 | 179 | 128 |
Datacrunch | A100 / 80 GB | 1 | 1.85 | 232 | 125 |
Crusoe Cloud | A100 / 80 GB | 4 | 6.600 | 821 | 124 |
Crusoe Cloud | A100 / 40 GB | 1 | 1.450 | 179 | 123 |
Crusoe Cloud | A100 / 80 GB | 2 | 3.300 | 406 | 123 |
Seeweb | RTX A6000 / 48 GB | 2 | 1.480 | 179 | 121 |
Vast.ai | A100 / 80 GB | 1 | 2.000 | 232 | 116 |
Lambda | A100 / 80 GB | 8 | 12 | 1,362 | 114 |
Datacrunch | A100 / 80 GB | 4 | 7.4 | 821 | 111 |
Datacrunch | A100 / 80 GB | 2 | 3.7 | 406 | 110 |
CoreWeave | A100 / 80 GB | 1 | 2.210 | 232 | 105 |
CoreWeave | A100 / 80 GB | 1 | 2.210 | 232 | 105 |
FluidStack | A100 / 80 GB | 1 | 2.210 | 232 | 105 |
FluidStack | A100 / 80 GB | 1 | 2.210 | 232 | 105 |
Seeweb | A100 / 80 GB | 1 | 2.220 | 232 | 104 |
Crusoe Cloud | A100 / 80 GB | 8 | 13.200 | 1,362 | 103 |
Vast.ai | A100 / 80 GB | 4 | 8.000 | 821 | 103 |
Vast.ai | A100 / 80 GB | 2 | 4.000 | 406 | 101 |
Datacrunch | A100 / 80 GB | 8 | 14.8 | 1,362 | 92 |
Seeweb | A100 / 80 GB | 4 | 8.880 | 821 | 92 |
Oblivus Cloud | A100 / 80 GB | 1 | 2.55 | 232 | 91 |
Seeweb | A100 / 80 GB | 2 | 4.440 | 406 | 91 |
RunPod | A100 / 80 GB | 8 | 15.12 | 1,362 | 90 |
Vultr | A100 / 80 GB | 1 | 2.604 | 232 | 89 |
CoreWeave | A100 / 40 GB | 1 | 2.060 | 179 | 87 |
CoreWeave | A100 / 40 GB | 1 | 2.060 | 179 | 87 |
Crusoe Cloud | A100 / 80 GB | 8 | 15.600 | 1,362 | 87 |
Oblivus Cloud | A100 / 80 GB | 2 | 5.1 | 406 | 80 |
Oblivus Cloud | A100 / 80 GB | 4 | 10.2 | 821 | 80 |
Vultr | A100 / 80 GB | 4 | 10.417 | 821 | 79 |
Vultr | A100 / 80 GB | 2 | 5.208 | 406 | 78 |
Latitude.sh | H100 (80 GB) | 8 | 35.2 | 2,693 | 77 |
CoreWeave | H100 / 80 GB | 1 | 4.250 | 322 | 76 |
FluidStack | H100 / 80 GB | 1 | 4.250 | 322 | 76 |
Latitude.sh | H100 (80 GB) | 4 | 17.6 | 1,321 | 75 |
Oblivus Cloud | A100 / 40 GB | 1 | 2.39 | 179 | 75 |
ACE Cloud | A100 / 80 GB | 1 | 3.110 | 232 | 74 |
Latitude.sh | H100 (80 GB) | 1 | 4.4 | 322 | 73 |
Paperspace | A100 / 80 GB | 1 | 3.18 | 232 | 73 |
FluidStack | H100 / 80 GB | 1 | 4.760 | 322 | 68 |
CoreWeave | H100 / 80 GB | 1 | 4.780 | 322 | 67 |
Oblivus Cloud | A100 / 80 GB | 8 | 20.4 | 1,362 | 67 |
Lambda | V100 / 16 GB | 8 | 4.4 | 289 | 66 |
ACE Cloud | A100 / 80 GB | 2 | 6.200 | 406 | 65 |
Oblivus Cloud | V100 / 16 GB | 1 | 0.65 | 42 | 65 |
Paperspace | A100 / 80 GB | 4 | 12.72 | 821 | 65 |
Vultr | A100 / 80 GB | 8 | 20.833 | 1,362 | 65 |
Paperspace | A100 / 80 GB | 2 | 6.36 | 406 | 64 |
Latitude.sh | A100 (80 GB) | 8 | 23.2 | 1,362 | 59 |
Oblivus Cloud | V100 / 16 GB | 2 | 1.3 | 77 | 59 |
Oblivus Cloud | V100 / 16 GB | 4 | 2.6 | 153 | 59 |
Paperspace | A100 / 40 GB | 1 | 3.09 | 179 | 58 |
Paperspace | A100 / 80 GB | 8 | 25.44 | 1,362 | 54 |
CoreWeave | V100 / 16 GB | 1 | 0.800 | 42 | 53 |
Cirrascale | A100 / 80 GB | 8 | 26.030 | 1,362 | 52 |
Exoscale | V100 / 16 GB | 4 | 3.32 | 153 | 46 |
ACE Cloud | A100 / 80 GB | 2 | 9.280 | 406 | 44 |
Vultr | H100 / 80 GB | 1 | 7.5 | 322 | 43 |
Datacrunch | V100 / 16 GB | 1 | 1 | 42 | 42 |
Datacrunch | V100 / 16 GB | 2 | 2 | 77 | 39 |
Datacrunch | V100 / 16 GB | 4 | 4 | 153 | 38 |
Exoscale | V100 / 16 GB | 2 | 2.01 | 77 | 38 |
Cirrascale | A100 / 80 GB | 4 | 22.960 | 821 | 36 |
Datacrunch | V100 / 16 GB | 8 | 8 | 289 | 36 |
Exoscale | V100 / 16 GB | 1 | 1.38 | 42 | 30 |
OVHcloud | V100 / 16 GB | 1 | 1.97 | 42 | 21 |
OVHcloud | V100 / 16 GB | 2 | 3.94 | 77 | 20 |
OVHcloud | V100 / 16 GB | 4 | 7.89 | 153 | 19 |
Paperspace | V100 / 16 GB | 1 | 2.3 | 42 | 18 |
Spot GPUs
Cloud | GPU Type / Memory* | # of GPUs | Spot | Throughput** | Throughput** / $*** |
---|---|---|---|---|---|
Azure | A100 / 80 GB | 1 | 0.76 | 232 | 303 |
Azure | A100 / 80 GB | 4 | 3.05 | 821 | 269 |
Azure | A100 / 80 GB | 2 | 1.53 | 406 | 266 |
Jarvislabs | A100 / 40 GB | 1 | 0.79 | 179 | 227 |
GCP | A100 / 40 GB | 1 | 1.62 | 179 | 111 |
AWS | V100 / 16 GB | 1 | 0.92 | 42 | 46 |
AWS | V100 / 16 GB | 4 | 3.67 | 153 | 42 |
Azure | V100 / 16 GB | 1 | 1.04 | 42 | 40 |
AWS | V100 / 16 GB | 8 | 7.34 | 289 | 39 |
Azure | V100 / 16 GB | 2 | 2.08 | 77 | 37 |
Azure | V100 / 16 GB | 4 | 4.16 | 153 | 37 |
In all these throughput per dollar tables:
- Not all possible configurations are listed, more commonly used, deep learning focused configurations are included.
- West or Central US regions were used where possible
- These are the list prices for each category, high volume buyers can possibly get better pricing
Finally, it is good to clarify what “spot” means. Spot resources are:
– Interruptible so users need to keep on recording their progress. For example, Amazon EC2 P3, which provides V100 32 GB, is one of the most frequently interrupted Amazon spot services.4
– Offered on a dynamic, market-driven basis. The price for these GPU resources can fluctuate based on supply and demand, and users typically bid on the available spot capacity. If a user’s bid is higher than the current spot price, their requested instances will run.
Cloud GPU availability in different clouds
Input the model that you want in the search box to identify all cloud providers that offer it:
Provider | GPU | Multi-GPU | On-demand $ per single GPU hour*** |
---|---|---|---|
AWS | M60 8 GB | 1, 2, 4x | 1.14 |
AWS | T4 16 GB | 1, 2, 4, 8x | 1.20 |
AWS | A10G 24 GB | 1, 4, 8x | 1.62 |
AWS | V100 16 GB | 1, 4, 8x | 3.06 |
AWS | V100 32 GB | 8x | 3.90**** |
AWS | A100 40 GB | 8x | 4.10**** |
AWS | A100 80 GB | 8x | 5.12**** |
Azure | K80 12 GB | 1, 2, 4x | 0.90 |
Azure | T4 16 GB | 1, 4x | 1.20 |
Azure | P40 24 GB | 1, 2, 4x | 2.07 |
Azure | P100 16 GB | 1, 2, 4x | 2.07 |
Azure | V100 32 GB | 8x | 2.75 |
Azure | V100 16 GB | 1, 2, 4x | 3.06 |
Azure | A100 40 GB | 8x | 3.40**** |
Azure | A100 80 GB | 1, 2, 4x | 3.67 |
Azure | A100 80 GB | 8x | 4.10**** |
GCP | T4 16 GB | 1, 2, 4x | 0.75 |
GCP | K80 12 GB | 1, 2, 4, 8x | 0.85 |
GCP | P4 8 GB | 1, 2, 4x | 1.00 |
GCP | P100 16 GB | 1, 2, 4x | 1.86 |
GCP | V100 16 GB | 1, 2, 4, 8x | 2.88 |
GCP | A100 40 GB | 1, 2, 4, 8, 16x | 3.67 |
OCI | A100 40 GB | 8x | 4.00 |
OCI | A100 80 GB | 8x | 3.05 |
OCI | A10 24 GB | 1,2,4x | 2.00 |
OCI | V100 16 GB | 1,2,4,8x | 2.95 |
OCI | P100 16 GB | 1,2x | 1.275 |
ACE Cloud | A2 (16 GB) | 1, 2x | 0.59 |
ACE Cloud | A30 (32 GB) | 1, 2x | 0.95 |
ACE Cloud | A100 (80 GB) | 1, 2x | 3.11 |
Alibaba Cloud | A100 80 GB | 8x | |
Cirrascale | A100 (80 GB) | 4, 8x | 5.74 |
Cirrascale | RTX A6000 (48 GB) | 8x | 1.12 |
Cirrascale | RTX A5000 (24 GB) | 8x | 0.51 |
Cirrascale | RTX A4000 (16 GB) | 8x | 0.34 |
Cirrascale | A40 (48 GB) | 8x | 1.44 |
Cirrascale | A30 (24 GB) | 8x | |
Cirrascale | V100 (32 GB) | 4, 8x | 1.92 |
Cirrascale | RTX 6000 (48GB) | 8x | 1.18 |
CoreWeave | H100 (80 GB) | 1x | 4.25 |
CoreWeave | A100 (80 GB) | 1x | 2.21 |
CoreWeave | A100 (40 GB) | 1x | 2.06 |
CoreWeave | V100 (16 GB) | 1x | 0.80 |
CoreWeave | A40 (48 GB) | 1x | 1.28 |
CoreWeave | RTX 6000 (48 GB) | 1x | 1.28 |
CoreWeave | RTX 5000 (24 GB) | 1x | 0.77 |
CoreWeave | RTX 4000 (16 GB) | 1x | 0.61 |
CoreWeave | Quadro RTX 5000 (16 GB) | 1x | 0.57 |
CoreWeave | Quadro RTX 4000 (8 GB) | 1x | 0.24 |
Crusoe Cloud | A6000 (48 GB) | 1, 2, 4, 8x | 0.92 |
Crusoe Cloud | A40 (48 GB) | 1, 2, 4, 8x | 1.10 |
Crusoe Cloud | A100 (80 GB) | 1, 2, 4, 8x | 1.45 |
Crusoe Cloud | H100 (80 GB) | 8x | |
FluidStack | H100 (80 GB) | 1x | 4.25 |
FluidStack | A100 (80 GB) | 1x | 2.21 |
Jarvis Labs | Quadro RTX 5000 16 GB | 1x | 0.49 |
Jarvis Labs | Quadro RTX 6000 24 GB | 1x | 0.99 |
Jarvis Labs | RTX A5000 24 GB | 1x | 1.29 |
Jarvis Labs | RTX A6000 48 GB | 1x | 1.79 |
Jarvis Labs | A100 40 GB | 1x | 2.39 |
Lambda Labs | Quadro RTX 6000 24 GB | 1, 2, 4x | 1.25 |
Lambda Labs | RTX A6000 48 GB | 1, 2, 4x | 1.45 |
Lambda Labs | V100 16 GB | 8x | 6.8 |
Latitude.sh | H100 (80 GB) | 1, 4, 8x | 4.40 |
Latitude.sh | A100 (80 GB) | 8x | 23.2 |
LeaderGPU | A100 (40 GB) | ||
LeaderGPU | A10 (24 GB) | ||
LeaderGPU | V100 (32 GB) | ||
Linode | Quadro RTX 6000 24 GB | 1, 2, 4x | 1.50 |
OVH | V100 32 GB | 1, 2, 4x | 1.99 |
OVH | V100 16 GB | 1, 2, 4x | 1.79 |
Paperspace | Quadro M4000 8 GB | 1x | 0.45 |
Paperspace | Quadro P4000 8 GB | 1, 2, 4x | 0.51 |
Paperspace | Quadro RTX 4000 8 GB | 1, 2, 4x | 0.56 |
Paperspace | RTX A4000 16 GB | 1, 2, 4x | 0.76 |
Paperspace | Quadro P5000 16 GB | 1, 2, 4x | 0.78 |
Paperspace | Quadro RTX 5000 16 GB | 1, 2, 4x | 0.82 |
Paperspace | Quadro P6000 24 GB | 1, 2, 4x | 1.10 |
Paperspace | RTX A5000 24 GB | 1, 2, 4x | 1.38 |
Paperspace | RTX A6000 48 GB | 1, 2, 4x | 1.89 |
Paperspace | V100 32 GB | 1, 2, 4x | 2.30 |
Paperspace | V100 16 GB | 1x | 2.30 |
Paperspace | A100 40 GB | 1x | 3.09 |
Paperspace | A100 80 GB | 1, 2, 4, 8x | 3.19 |
Seeweb | RTXA6000 (48 GB) | 1, 2, 3, 4, 5x | 0.74 |
Seeweb | RTXA6000 (24 GB) | 1, 2, 3, 4, 5x | 0.64 |
Seeweb | A30 (24 GB) | 1, 2, 3, 4, 5x | 0.64 |
Seeweb | L4 (24 GB) | 1, 2, 3, 4, 5x | 0.38 |
Seeweb | A100 (80 GB) | 1, 2, 3, 4, 5x | 2.22 |
TensorDock | A100 (80 GB) | 1x | 1.40 |
TensorDock | L40 (40 GB) | 1x | 1.05 |
TensorDock | V100 (16 GB) | 1x | 0.22 |
TensorDock | A6000 (48 GB) | 1x | 0.47 |
TensorDock | A40 (48 GB) | 1x | 0.47 |
TensorDock | A5000 (24 GB) | 1x | 0.21 |
TensorDock | A4000 (16 GB) | 1x | 0.13 |
TensorDock | RTX 4090 (24 GB) | 1x | 0.37 |
TensorDock | RTX 3090 (24 GB) | 1x | 0.22 |
TensorDock | RTX 3080 Ti (12 GB) | 1x | 0.17 |
TensorDock | RTX 3080 (10 GB) | 1x | 0.17 |
TensorDock | RTX 3070 Ti (8 GB) | 1x | 0.14 |
TensorDock | RTX 3060 Ti (8 GB) | 1x | 0.10 |
TensorDock | RTX 3060 (12 GB) | 1x | 0.10 |
Vast.ai | L40 (45 GB) | 1, 2, 4x | 1.10 |
Vast.ai | A100 (40 GB) | 1, 2, 4x | 1.40 |
Vast.ai | A40 (48 GB) | 1, 2x | 0.40 |
Vast.ai | A6000 (24 GB) | 1, 2, 4, 8x | 0.44 |
Vast.ai | A5000 (24 GB) | 1, 2, 4, 8x | 0.20 |
Vast.ai | A4000 (16 GB) | 1, 2, 4, 5, 8x | 0.15 |
Vast.ai | V100 (16 GB) | 2, 5x | 0.18 |
Voltage Park | H100 80 GB | 8x | 1.89**** |
Vultr | L40S 48 GB | 1, 2, 4, 8x | 1.75 |
Vultr | H100 80 GB | 1x | 7.50 |
Vultr | A100 80 GB | 1, 2, 4, 8x | 2.60 |
Vultr | A40 (48 GB) | 1, 4x | 1.83 |
Vultr | A16 (16 GB) | 1, 2, 4, 8, 16x | 0.51 |
**** Computed values. This was needed when single GPU instances were not available.5 6
Other cloud GPU considerations
Availability: Not all GPUs listed above may be available due to capacity constraints of the cloud providers and increasing demand for generative AI.
Data security: For example, cloud GPU marketplaces like Vast.ai offer significantly lower prices but depending on the specific resource requested, the data security of the workload could be impacted, givings hosts the capability to access workloads. Since we prioritized enterprise GPU needs, Vast.ai wasn’t included in this benchmark.
Ease of use: Documentation quality is a subjective metric but developers prefer some cloud providers’ documentation over others. In this discussion, GCP’s documentation was mentioned as lower quality than other tech giants’.7
Familiarity: Even though cloud providers put significant effort into making their services easy-to-use, there is a learning curve. That is why major cloud providers have certifications systems in place. Therefore, for small workloads, the cost savings of using a low cost provider may be less than the opportunity cost of the time it takes a developer to learn how to use their cloud GPU offering.
Buy GPUs or rent cloud GPUs
Buying makes sense
– If your company has the know-how and preference to host the servers or manage colocated servers.
– For uninterruptible workloads: For the volume of GPUs for which you can ensure a high utilization (e.g. more than 80%) for a year or more.8
– For interruptible workloads: The high utilization period quoted above needs to be a few times longer since on-demand (uninterruptible computing) prices tends to be a few times more expensive than spot (interruptible computing) prices.
Our recommendation for businesses with heavy GPU workloads is a mix of owned and rented GPUs where guaranteed demand runs on owned GPUs and variable demand runs on the cloud. This is why tech giants like Facebook are building their own GPU clusters including hundreds of GPUs.9
Buyers may be tempted to consider consumer GPUs which offer a better price/performance ratio however, the EULA of their software prohibits their use in data centers.10 Therefore, they are not a good fit for machine learning except for minor testing workloads on data scientists’ machines.
What are the top cloud GPU hardware?
Almost all cloud GPUs use Nvidia GPU instances. AMD and other providers also offer GPUs however due to various reasons (e.g. limited developer adoption, lower price per performance etc.), their GPUs are not as widely demanded as Nvidia GPUs.
To see cloud GPU providers that offer non-Nvidia GPUs, please check out the comprehensive list of cloud GPU providers.
Read about all AI chips / hardware.
What are the top cloud GPU platforms?
Top cloud GPU providers are:
- AWS
- Microsoft Azure
- CoreWeave
- Google Cloud Platform (GCP)
- IBM Cloud
- Jarvis Labs
- Lambda Labs
- NVIDIA DGX Cloud
- Oracle Cloud Infrastructure (OCI)
- Paperspace CORE
For more on these providers, check out cloud gpu providers.
If you are not sure about cloud GPUs, explore other options like serverless GPU.
For tools allow users to collaborate on, check out cloud collaboration tools.
If you are unclear about what cloud GPUs are, here is more context:
What is a cloud GPU?
Unlike a CPU, which may have a relatively small number of cores optimized for sequential serial processing, a GPU can have hundreds or even thousands of smaller cores designed for multi-threading and handling parallel processing workloads.
A cloud GPU refers to a certain way of GPU usage that’s provided as a service through cloud computing platforms. Much like traditional cloud services, a cloud gpu allows you to access high-performance computing resources spot or on-demand, without the need for upfront capital investment in hardware.
What are the functions/application areas of cloud GPUs?
Cloud GPUs are primarily used for processing tasks that require high computational power. Here are some of the primary uses for cloud GPUs:
Machine Learning and AI
GPUs are particularly effective at handling the complex calculations required for machine learning (ML) and artificial intelligence (AI) models. They can process multiple computations in parallel, making them suitable for training large neural networks and algorithms.
Deep learning
Deep learning is a sub-field of machine learning. Deep learning algorithms greatly benefit from the parallel processing capabilities of GPUs, making training and inference faster and more efficient.
Data processing
Data analysis
GPUs are used to accelerate computing and data processing tasks, such as Big Data analysis and real-time analytics. They can handle high-throughput, parallel processing tasks more efficiently than CPUs.
Scientific computing
In scientific research, cloud GPUs can handle computations for simulations, bioinformatics, quantum chemistry, weather modeling, and more.
Simulations
Certain complex simulations can run more efficiently on GPUs.
Gaming & entertainment
Cloud GPUs are used to provide cloud gaming services, such as Google’s Stadia or NVIDIA’s GeForce Now, where the game runs on a server in the cloud, and the rendered frames are streamed to the player’s device. This allows high-quality gaming without the need for a powerful local machine.
Graphics rendering
GPUs were initially designed to handle computer graphics, and they still excel in this area. Cloud GPUs are used for 3D modeling and rendering, 3D visualizations, virtual reality (VR), computer-aided design (CAD), and computer-generated imagery (CGI).
Video processing
They’re used in video encoding and decoding, video editing, color correction, effects rendering, and other video processing tasks.
Cryptocurrency mining
GPUs are also used in tasks like cryptocurrency mining. However application-specific integrated circuits (ASICs) are offering better economics for more commonly mined crypto currencies.
Notes
Cloud providers are constantly updating their offering, part of this research could be outdated.
AIMultiple’s research is sponsored by cloud GPU providers. Sponsors can be identified by links to their websites and they are ranked at the top of lists that they participate in.
External links
- 1. “Deep Learning GPU Benchmarks“, Lambda Labs, Retrieved July 15, 2023
- 2. “Open LLM-Perf Leaderboard“, Hugging Face, Retrieved July 15, 2023
- 3. “the-full-stack/website/docs/cloud-gpus“, GitHub, Retrieved July 15, 2023.
- 4. “Spot Instance advisor“, Amazon, Retrieved July 19, 2023
- 5. “The Ultimate Guide to Cloud GPU Providers“, Paperspace, Retrieved July 15, 2023
- 6. “CloudOptimizer“, CloudOptimizer, Retrieved July 15, 2023
- 7. “Cloud GPU Resources and Pricing“, Hacker News, Retrieved July 16, 2023
- 8. “Cloud GPU Resources and Pricing“, Hacker News, Retrieved July 16, 2023
- 9. “Meta Works with NVIDIA to Build Massive AI Research Supercomputer“, Nvidia, Retrieved July 16, 2023
- 10. “License For Customer Use of NVIDIA GeForce Software“, Nvidia, Retrieved July 16, 2023
Cem has been the principal analyst at AIMultiple since 2017. AIMultiple informs hundreds of thousands of businesses (as per similarWeb) including 60% of Fortune 500 every month.
Cem's work has been cited by leading global publications including Business Insider, Forbes, Washington Post, global firms like Deloitte, HPE, NGOs like World Economic Forum and supranational organizations like European Commission. You can see more reputable companies and media that referenced AIMultiple.
Throughout his career, Cem served as a tech consultant, tech buyer and tech entrepreneur. He advised businesses on their enterprise software, automation, cloud, AI / ML and other technology related decisions at McKinsey & Company and Altman Solon for more than a decade. He also published a McKinsey report on digitalization.
He led technology strategy and procurement of a telco while reporting to the CEO. He has also led commercial growth of deep tech company Hypatos that reached a 7 digit annual recurring revenue and a 9 digit valuation from 0 within 2 years. Cem's work in Hypatos was covered by leading technology publications like TechCrunch and Business Insider.
Cem regularly speaks at international technology conferences. He graduated from Bogazici University as a computer engineer and holds an MBA from Columbia Business School.
To stay up-to-date on B2B tech & accelerate your enterprise:
Follow on
Comments
Your email address will not be published. All fields are required.