Scan code: LN111283 Manufacturer code: TCSA100M-EDU

40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY

40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY

Thumbnail 1 : 40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY Thumbnail 2 : 40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY Thumbnail 3 : 40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY Thumbnail 4 : 40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY
Scan code: LN111283 Manufacturer code: TCSA100M-EDU
Delivery Options
  • By DPD to your specified address. | £11.50 Receive SMS with one-hour delivery window Weekend, timed and European delivery options are available at checkout

    Have your parcel delivered by DPD to your specified address. Receive SMS with one-hour delivery window

    Weekend, timed and European delivery options are available at checkout
  • Collect instore Collect from our Bolton store, BL6 6PE | Free

    Order online, collect from our Bolton store (25-28 Enterprise Park, Middlebrook, Horwich, Bolton, BL6 6PE)

Services
  • 48HR REPLACEMENT If you need to return this item, your replacement will be dispatched within 2 working days of your product arriving back at Scan. More info
Product Overview

This PNY A100 card is available for educational institutions ONLY. If you are educational institution interested in this product, please contact the SCAN Managed Account Team by telephone to discuss your requirements. If you do not meet the requirement of being an educational institution & are still interested in purchasing the A100 graphics card, PLEASE CLICK HERE FOR THE REGULAR RETAIL CARD.


Unprecedented Acceleration at Every Scale

PNY NVIDIA A100 Tensor Core GPU provides users with supreme acceleration and power. The NVIDIA A100 Tensor Core GPU delivers unprecedented acceleration at every scale for AI, data analytics, and HPC to tackle the world’s toughest computing challenges. As the engine of the NVIDIA data center platform, A100 can efficiently scale up to thousands of GPUs or, using new Multi-Instance GPU (MIG) technology, can be partitioned into seven isolated GPU instances to accelerate workloads of all sizes. A100’s third-generation Tensor Core technology now accelerates more levels of precision for diverse workloads, speeding time to insight as well as time to market. The NVIDIA Ampere architecture, designed for the age of elastic computing, delivers the next giant leap by providing unmatched acceleration at every scale, enabling these innovators to do their life’s work.


NVIDIA Ampere Architecture

A100 accelerates workloads big and small. Whether using MIG to partition an A100 GPU into smaller instances, or NVLink to connect multiple GPUs to accelerate large-scale workloads, A100 can readily handle different-sized acceleration needs, from the smallest job to the biggest multi-node workload. A100’s versatility means IT managers can maximize the utility of every GPU in their data centre around the clock.


Third-Generation Tensor Cores

A100 delivers 312 teraFLOPS (TFLOPS) of deep learning performance. That’s 20X Tensor FLOPS for deep learning training and 20X Tensor TOPS for deep learning inference compared to NVIDIA Volta™ GPUs.


Next-Generation NVLink

NVIDIA NVLink in A100 delivers 2X higher throughput compared to the previous generation. When combined with NVIDIA NVSwitch™, up to 16 A100 GPUs can be interconnected at up to 600 gigabytes per second (GB/sec) to unleash the highest application performance possible on a single server. NVLink is available in A100 SXM GPUs via HGX A100 server boards and in PCIe GPUs via an NVLink Bridge for up to 2 GPUs.


Multi-Instance GPU (MIG)

An A100 GPU can be partitioned into as many as seven GPU instances, fully isolated at the hardware level with their own high-bandwidth memory, cache, and compute cores. MIG gives developers access to breakthrough acceleration for all their applications, and IT administrators can offer right-sized GPU acceleration for every job, optimizing utilization and expanding access to every user and application.


HBM2

With 40 gigabytes (GB) of high-bandwidth memory (HBM2), A100 delivers improved raw bandwidth of 1.6TB/sec, as well as higher dynamic random-access memory (DRAM) utilization efficiency at 95 percent. A100 delivers 1.7X higher memory bandwidth over the previous generation.


Structural Sparsity

AI networks are big, having millions to billions of parameters. Not all of these parameters are needed for accurate predictions, and some can be converted to zeros to make the models “sparse” without compromising accuracy. Tensor Cores in A100 can provide up to 2X higher performance for sparse models. While the sparsity feature more readily benefits AI inference, it can also improve the performance of model training.


Up to 7x Higher Performance with Multi Instance GPU for Ai Inference

The NVIDIA A100 Tensor Core GPU is the flagship product of the NVIDIA data centre platform for deep learning, HPC, and data analytics. The platform accelerates over 700 HPC applications and every major deep learning framework. It’s available everywhere, from desktops to servers to cloud services, delivering both dramatic performance gains and cost-saving opportunities.

Features
Architecture: Ampere.
Memory: 40GB HBM2.
Memory Bandwidth: 1.6TB/s.
Interface: PCIe 4.0.
Power: 250W.
Specifications
Key Specifications
Graphics Chipset NVIDIA A100
Edition A100 PCIe
Manufacturing process  
Microarchitecture Ampere
GPU Name  
Cores & Clocks
NVIDIA CUDA Cores  
Core Clock  
Boost Clock  
Video Memory (VRAM)
Memory Size 40 GB
Memory Type HBM2
Memory Clock  
Memory Bit Rate  
Memory Bandwidth 1600 GB/s
ECC Technology No
Cooling
Cooling Passive
I/O & Connectivity
Interface PCIe 4.0 (x16)
Interface Bandwidth  
Graphics Output  
Visuals
Multi GPU Support NVLink
NVLink Support Upto 16x A100 GPUs via NVLink
Total NVLink Bandwidth 600 GB/s
Microsoft DirectX Support  
HDCP Ready No
Multi Monitor Support  
Concurent Users N/A
H.264 1080p30 Streams N/A
Maximum Digital Resolution  
Maximum VGA Resolution  
Supported Graphics APIs  
Compute Performance
Supported Compute APIs
  • CUDA
  • DirectCompute
  • OpenACC
  • OpenCL
Single Precision (FP32) Processing Yes
Single Precision (FP32) Performance 1905 teraFLOPS
Double Precision (FP64) Processing Yes
Double Precision (FP64) Performance 9.7 teraFLOPS
Tensor Processing Yes
Deep Learning (Tensor) Performance 312 teraFLOPS
NVIDIA Tensor Cores  
Integer Operations (INT8)  
Ray-Tracing
NVIDIA RT Cores  
RTX-OPS  
Rays Cast  
Power & Thermals
Graphics Card Power Connectors  
Board Power 250 W
Minimum Recommended PSU  
Maximum GPU Temperature  
Physical
Form Factor  
Low Profile Compatible No
Low Profile Support  
Dimensions  
Package Type  
Additional Information
Scan Code LN111283
Model Number TCSA100M-EDU
GTIN 3536403378035
Warranty

Please note your statutory rights are not affected.

For further information regarding Scan's warranty procedure please see our terms and conditions

Details
Duration:
24 months
Type:
Direct
Manufacturer Contact Details
Manufacturer:
www.pny.com
TekSpek Guides
Tesla vs GeForce
Tesla vs GeForce
Date Issued: 2nd Dec 2010

NVIDIA® CUDA parallel computing architecture is enabled on GeForce, Quadro, and Tesla products. Whereas GeForce and Quadro are designed for consumer graphics and professional visualization respectively, the NVIDIA® Tesla™ product family is designed ground-up for parallel computing and offers exclusive computing features.

Read More

Graphics Card Outputs
Graphics Card Outputs
Date Issued: 8th Oct 2010

Modern desktop computers and notebooks comprise of a CPU, motherboard, graphics, storage, and, usually an optical drive. Computers have a number of ports and sockets that enable the user to plug-in various peripherals such as a printer, USB mouse, or, perhaps most importantly of all, an Internet connection.

Read More

CUDA Technology
CUDA Technology
Date Issued: 22nd Oct 2008

CUDA technology is the world’s only C language environment that enables programmers and developers to write software to solve complex computational problems in a fraction of the time by tapping into the many-core parallel processing power of GPUs.

Read More

Overclocking
Overclocking
Date Issued: 20th Oct 2008

This TekSpek explains why you’d want to overclock your graphics board, the risks in doing so and how you can go about doing it.

Read More

GPU Antialiasing
GPU Antialiasing
Date Issued: 20th Oct 2008

This TekSpek will assume you know the affects of applying a level of anti-aliasing (AA) on your 3D accelerator, be it via the driver control panel or via a control in your game. We assume you know the effect it has on image quality, so you can think about a before and after scenario. So this TekSpek isn’t about explaining what it does as such, although it will, it’s about explaining the how and why.

Read More

GPU
GPU
Date Issued: 20th Oct 2008

Explaining how a modern GPU works in completeness would take a book. Or two. Per class of chip. Per vendor. They’re extraordinarily complex pieces of engineering and production, and the end result contains more transistors than multiple modern x86 processors.

Read More

DirectX 10
DirectX 10
Date Issued: 25th Jun 2008

This TekSpek explains DirectX 10, which graphics cards support it, how they work and what the consumer’s choices are.

Read More

ESD
ESD
Date Issued: 14th Jun 2008

We’ve all been victims of static electricity at some point. Perhaps somebody’s used a balloon to make your hair stand on end, or you’ve walked across the office and been ‘shocked’ by a metal door knob? That’s electrostatic discharge at work.

Read More

40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY
40GB PNY NVIDIA A100 PCI-e, Ampere Graphics Card For Education ONLY