Pricing - In $USD
100% Predictable Pricing Guarantee
GPUs On TIR Platform
Spot Instances
GPU Dedicated Compute with HGX H100
NVIDIA HGX H100 combines the power for H100 GPUs with high-speed interconnects, thus forming one of the most powerful servers. It can host up to eight H100 Tensor Core GPUs and four third-generation NVSwitch. Each GPU has several fourth generation NVLink ports and connects to all four NVSwitches. Thus, it supports configurations with up to eight GPUs, providing 640 GB of GPU memory with an aggregate memory bandwidth of 24 terabytes per second. With an astonishing 32 petaFLOPS of performance, it stands as the most powerful accelerated scale-up server platform for AI and HPC.
GPU Dedicated Compute with A100
GPU Dedicated Compute with L40S
The NVIDIA L40S GPU, based on the Ada Lovelace architecture is a full height, full-length (FHFL), and delivers end to end acceleration for AI enabled applications such as Generative AI, Model training and inference, and video applications.
GPU Dedicated Compute with A100 40GB
The NVIDIA A100 40GB is a high-performance data center GPU designed for deep learning, AI, and HPC workloads. With its advanced architecture and large memory capacity, the A100 40GB can accelerate a wide range of compute-intensive applications, including training and inference for natural language processing, image recognition, and more.
GPU Dedicated Compute with A40
The NVIDIA A40 GPU is an evolutionary leap in performance and multi-workload capabilities from the data center, combining best-in-class professional graphics with powerful compute and AI acceleration to meet today’s design, creative, and scientific challenges.
Driving the next generation of virtual workstations and server-based workloads, NVIDIA A40 brings state-of-the-art features for ray-traced rendering, simulation, virtual production, and more to professionals anytime, anywhere.
GPU Dedicated Compute with A30
AI Inference and Mainstream Compute for Every Enterprise Bring accelerated performance to every enterprise workload with NVIDIA A30 Tensor Core GPUs.
With NVIDIA Ampere architecture Tensor Cores, it delivers speedups securely across diverse workloads, including AI inference at scale and high-performance computing (HPC) applications.
By combining fast memory bandwidth and low-power consumption in a PCIe form factor—optimal for mainstream servers—A30 enables an elastic data center and delivers maximum value for enterprises.
GPU Dedicated Compute with L4
Equipped with cutting-edge fourth-generation Tensor Cores and an impressive 1.5X larger GPU memory, the NVIDIA L4 GPU, in combination with the CV-CUDA® library, takes video content understanding to unprecedented heights.
GPU Dedicated Compute with V100
Startups, SMEs, and enterprises across various sectors use E2E Cloud for their application workloads. Now use Tesla GPUs on E2E Cloud, to power your cutting-edge, High performance applications and to run your research & development workloads.
Cloud GPUs on GPU Smart Dedicated
Startups, SMEs, and enterprises across various sectors use E2E Cloud for their application workloads. Now use Tesla GPUs on E2E Cloud, to power your cutting-edge, High performance applications and to run your research & development workloads.
E2E Object Storage
Also known as EOS, E2E object storage is an SSD-based S3-compatible object storage service designed for demanding workloads like machine learning and deep learning.