skip to Main Content
Applied Data Systems is your NVIDIA Headquarters for Everything NVIDIA

DGX A100

NVIDIA DGX A100 Game Changing Performance

The NVIDIA DGX A100 Accelerated Compute Server delivers unprecedented performance for deep learning training and inference. Organizations can now deploy data-intensive, deep learning frameworks with confidence. DGX A100 enables the cutting-edge DL/ML and AI innovation data scientists desire, with the dependability IT requires. Available in POD design such as AgilityFlexAI with Spectrum Scale Storage Solution. Now available with 80GB of HBM2e memory with over 2TB’s per second of memory bandwidth.

NVIDIA DGX Station A100

NVIDIA DGX Station A100 brings AI supercomputing to data
science teams, offering data center technology without a data
center or additional IT infrastructure. Designed for multiple,
simultaneous users, DGX Station A100 leverages server-grade
components in an office-friendly form factor. It’s the only
system with four fully interconnected and Multi-Instance
GPU (MIG)-capable NVIDIA A100 Tensor Core GPUs with up
to 320 GB of total GPU memory that can plug into a standard
power outlet, resulting in a powerful AI appliance that you can
place anywhere.


NVIDIA A100 Tensor Core GPU

NVIDIA A100’s third-generation Tensor Cores with Tensor Float (TF32) precision provide up to 20X higher performance over the prior generation with zero code changes and an additional 2X boost with automatic mixed precision and FP16. When combined with third-generation NVIDIA® NVLink®, NVIDIA NVSwitch, PCI Gen4, NVIDIA Mellanox InfiniBand, and the NVIDIA Magnum IO software SDK, it’s possible to scale to thousands of A100 GPUs. This means that large AI models like BERT can be trained in just 37 minutes on a cluster of 1,024 A100s, offering unprecedented performance and scalability.

Available in PCIe Gen4 or SXM4 form factor in our rackmount servers.

NVIDIA Mellanox Ethernet and Infiniband Networking Solutions

The seventh generation of the NVIDIA® Mellanox® InfiniBand architecture, featuring NDR 400 Gb/s InfiniBand, gives AI developers and scientific researchers the fastest networking performance available to take on the world’s most challenging problems. NVIDIA Mellanox Infiniband® is paving the way with software-defined networking, In-Network Computing acceleration, remote direct-memory access (RDMA), and the fastest speeds and feeds—including impressive advancements over the previous HDR InfiniBand generation.

NVIDIA DGX SuperPOD Solution for Enterprise

NVIDIA DGX SuperPOD Solution for Enterprise incorporates the best practices and know-how gained from the world’s largest AI deployments, designed to solve the most challenging AI opportunities facing organizations. For enterprises that need a trusted and turnkey approach to AI innovation at scale, we’ve taken our industry-leading reference architecture and wrapped it in a comprehensive solution and services offering. NVIDIA DGX SuperPOD Solution for Enterprise delivers a full-service experience that delivers industry-proven results in weeks instead of months to every organization that needs leadership-class infrastructure, with a white-glove implementation that’s intelligently integrated with your business, so your team can deliver results sooner.


Back To Top