skip to Main Content
1-844-371-4949 info@applieddatasystems.com

NVIDIA GeForce RTX 3090

The GeForce RTX™ 3090 is a big ferocious GPU (BFGPU) with TITAN class performance. It’s powered by Ampere—NVIDIA’s 2nd gen RTX architecture—doubling down on ray tracing and AI performance with enhanced Ray Tracing (RT) Cores, Tensor Cores, and new streaming multiprocessors. Plus, it features a staggering 24 GB of G6X memory, all to deliver the ultimate gaming experience. Available in our custom workstations now.

NVIDIA RTX A6000

GPU Memory: 48 GB GDDR6 with error-correcting code (ECC)
Display Ports: 4x DisplayPort 1.4*
Max Power Consumption: 300 W
Graphics Bus: PCI Express Gen 4 x 16
Form Factor: 4.4” (H) x 10.5” (L) dual slot
Thermal: Active
NVLink: 2-way low profile (2-slot and 3-slot bridges) Connect 2 RTX A6000
vGPU Software Support: NVIDIA GRID®, NVIDIA Quadro® Virtual Data Center Workstation,
NVIDIA Virtual Compute Server
vGPU Profiles Supported: 1 GB, 2 GB, 3 GB, 4 GB, 6 GB, 8 GB, 12 GB, 16 GB, 24 GB, 48 GB
VR Ready: Yes

Available in our line of custom workstations and rackmount servers.

NVIDIA A100 Tensor Core GPU

NVIDIA A100’s third-generation Tensor Cores with Tensor Float (TF32) precision provide up to 20X higher performance over the prior generation with zero code changes and an additional 2X boost with automatic mixed precision and FP16. When combined with third-generation NVIDIA® NVLink®, NVIDIA NVSwitch, PCI Gen4, NVIDIA Mellanox InfiniBand, and the NVIDIA Magnum IO software SDK, it’s possible to scale to thousands of A100 GPUs. This means that large AI models like BERT can be trained in just 37 minutes on a cluster of 1,024 A100s, offering unprecedented performance and scalability.

Available in PCIe Gen4 or SXM4 form factor in our rackmount servers.

Highlights of NVIDIA A100

Deep Learning Training:

NVIDIA A100’s third-generation Tensor Cores with Tensor Float (TF32) precision provide up to 20X higher performance over the prior generation with zero code changes and an additional 2X boost with automatic mixed precision and FP16

Deep Learning Inference:

A100 introduces groundbreaking new features to optimize inference workloads. It brings unprecedented versatility by accelerating a full range of precisions, from FP32 to FP16 to INT8 and all the way down to INT4. Multi-Instance GPU (MIG) technology allows multiple networks to operate simultaneously on a single A100 GPU for optimal utilization of compute resources. And structural sparsity support delivers up to 2X more performance on top of A100’s other inference performance gains.

High-Performance Computing

A100 introduces double-precision Tensor Cores, providing the biggest milestone since the introduction of double-precision computing in GPUs for HPC. This enables researchers to reduce a 10-hour, double-precision simulation running on NVIDIA V100 Tensor Core GPUs to just four hours on A100. HPC applications can also leverage TF32 precision in A100’s Tensor Cores to achieve up to 10X higher throughput for single-precision dense matrix multiply operations.

NVIDIA GPUs

Back To Top