AI Development Hardware Buyers Guide

Development is the initial stage of any AI journey, prior to training and inferencing. As such it is seen as the most critical phase, as your development foundations set the scene for all future model training, fine-tuning and scaling of your projects. Mistakes made at the development phase can be very costly further down the line in terms of both expenditure and time, thus hardware selection to carry out development tasks is crucial to get right.

This guide takes you through the various AI development hardware options, explaining their differences and suitability for projects including small language models (SLMs), large language models (LLMs), generative AI, agentic AI and physical AI models.

Starting your AI Project

Although model development is the first of the three classic stages of an AI project, this phase is always preceded by a number of equally critical steps, including high-level planning, project scope setting and data preparation, as illustrated below.

Problem Statement

Project scope and high level ROI

Data Preparation

Classification, cleaning and structure

Model Development

Education and resource allocation

Model Training

Optimisation and scaling

Model Integration

Inferencing and deployment

Governance

Maintenance and compliance

It may not be immediately obvious, but ensuring your project scope is realistic and achievable has a large impact on what AI development hardware you’ll ultimately require. Similarly, undertaking adequate data preparation, including classification, structuring and cleansing of data, ensures you know the precise size of datasets you’ll be dealing with - thus providing an indication for GPU sizing, memory capacity and storage choices. Perhaps most crucially, data size also impacts the ability to scale your model as you iterate numerous options and evaluate outcomes. This data preparation stage can be up to as much as 60% of the entire time of your entire AI project, indicating how important it is to get it right. You can learn more about these crucial pre-development stages by reading our A-Z of AI White Paper.

AI Model Development

When it comes to using your prepared data in model development, chances are your project isn’t unique, so there may well be pre-trained frameworks or optimised foundation models (FMs). These save time and effort when building an AI pipeline, and removing the need for much of the subsequent development work. A prime example is NVIDIA AI Enterprise (NVAIE), which features in excess of 30 distinct and interlinked pre-trained frameworks, optimised for NVIDIA GPUs, designed for end-to-end implementation of AI projects such as medical imaging, autonomous vehicles, avatars, drug discovery, robotics, generative AI and many more.

Understanding the frameworks or FMs available to you may also have a significant impact on your AI development platform choice, as it may dictate a minimum GPU memory size or level of performance required to complete this phase within a given timeframe. When it comes to model size, some FMs may already number into the several billion-parameter space, before you start fine-tuning them into your own tailored (often even larger) model. It is therefore key to understand the relative size of models, how they might scale and the GPU hardware that will be capable of handling this. Just for context, between the 1950s and to 2018, AI model size grew by seven orders of magnitude (from 000’s to 30M) – yet from 2018 to 2022 alone, it has grown another four orders of magnitude (from 30M to 20B). Today 400B parameter models are not uncommon and ChatGPT-4 is rumoured to have 1.8T parameters.

AI Model Development

Mike and the CreativeBloke team feel that cloud-based workflows are invaluable for creative professionals seeking high-performance solutions without being tied to a fixed workspace. Freelancers, studios, and remote teams alike can benefit from its scalability and sustainability. As a next step, Mike intends to integrate AI-powered tools and explore collaborative opportunities by connecting multiple remote systems into a unified pipeline. The presumption is that this would further streamline workflows while enabling even greater creative freedom.

Type of AI Model Typical Use Case Parameters Dataset Size Typical GPU(s) Required
SLMs Limited scope chatbots / NLP / On-device applications 100M - 7B 1 - 8GB Consumer GPU(s) in laptop / desktop
LLMs Content creation / advanced chatbots / Real-time translation 7B - 20B 10 - 15GB Professional GPU(s) in workstation
Generative AI Advanced content creation / Personalised experiences / Drug Discovery 20B - 70B 20 - 40GB Multiple professional GPUs in workstation
Agentic AI Personalised interactions / Data-driven insights / Autonomous vehicles 70B - 200B 60-150GB Multiple professional GPUs in servers / cluster
Physical AI Robotics / Digital Twins 200B - xT 200-750GB Multiple professional GPU servers in cluster

It is worth clarifying that this table is for guidance only - absolute size (in GB) of any model is determined by number of parameters and the size of each parameter. Similarly, there will be small agentic AI models if their function is very focused and there will be very large LLMs if translation of many languages is the goal. It is also worth pointing out that the dataset size mentioned is the likely final size, so you need to consider capacity for numerous versions and many many iterations before reaching the final model. You can find more in-depth information about model development by reading our A-Z of AI White Paper.

AI Hardware

The following development options are compared and discussed in light of the model sizes above, offering recommendations and advice about the most suitable option for various scenarios. However, as previously stated thorough planning and scoping phases will lead to much more accurate provisional model sizes and better insight into hardware choice.

Traditional AI workstations are built around the industry-norm of an x86 CPU (either AMD or Intel) plus one or more NVIDIA RTX GPUs connected via the PCIe bus. The major benefit of this approach is that as a well-established system architecture over decades, such systems are extremely cost-effective and very easy to expand with more powerful processors, memory, storage and networking as your needs grow.

The downsides are that the CPU and GPU have separate pools of memory and that NVIDIA RTX GPUs are currently limited to 96GB of memory, so require multiple cards to run generative AI and reasoning AI models effectively. A new approach, using system-on-chip (SoC) provides a single CPU/GPU resource connected to a large combined memory pool.

Click the tabs below to explore these options, or of course you can contact our AI team for more information or advice.

NVIDIA DGX Spark

Nvidia DGX Spark

The DGX Spark is NVIDIA’s latest appliance for developers looking for a desktop solution to develop and fine-tune generative AI and reasoning AI models. These are particularly challenging workloads as they have a large memory footprint, so are too large to run on many existing GPUs. The DGX Spark has been built from the ground-up to provide a single large memory pool. It achieves this by replacing the old discrete CPU and GPU paradigm with a SoC, combining CPU and GPU cores together in a single unit, making the unit very compact despite its considerable performance.

Architecture

Known as the GB10 Grace Blackwell Superchip, this SoC comprises a Blackwell GPU and 20 Arm CPU cores, sharing a unified 128GB of memory. The GPU element of the SoC is equipped with the latest 5th gen Tensor cores, which are optimised for performing FP4 calculations, the most commonly used precision level for generative AI and reasoning AI models, while the CPU element has ten Cortex X925 and ten Cortex A725 cores.

Diagram of internal components
Connectivity
Wi-Fi, Bluetooth, USB
ConnectX
NCCL, RDMA, GPUDirect
4 TB SSD
GB10 Superchip Blackwell GPU
1 PetaFLOP FP4 AI Compute
Grace GPU
20 Arm Cores
High Bandwidth Unified Memory
128 GB Low Power DDR5X

The SoC is supported by 4TB of NVMe SSD storage, alongside a 10GbE Ethernet port, ConnectX-7 SmartNIC and WiFi 7 for ingesting data. Just like a typical desktop PC, the DGX Spark has a HDMI port to connect a monitor, plus four USB Type C ports to connect peripherals. Furthermore, two DGX Spark units can also be clustered using a special LinkX cable, effectively creating a combined 256GB memory pool for running extremely larger models.

The DGX Spark runs DGX OS, which is a NVIDIA-customised version of Ubuntu Linux. This enables developers to unleash the full potential of their DGX Spark with a proven software platform built around a complete AI software stack including access to NVIDIA NIM microservices, Blueprints and AI Enterprise. This is of course the same software stack as other NVIDIA AI appliances, making it easy to transfer your newly-developed AI model onto more powerful MGX, HGX or DGX systems, whether they’re in your own datacentre or in the cloud.

Relative Performance & Capability

The DGX Spark has an FP4 performance of 1,000 TOPS, which as you can see in the table below, isn’t the highest - however with a GPU memory capacity of 128GB (or 256GB if you combine two Sparks), it clearly has the best potential for natively and cost-effectively developing extremely large AI models.

System NVIDIA DGX Spark 3XS AI Laptop with GeForce RTX 5090 3XS AI Workstation with GeForce RTX 5090 3XS AI Workstation with RTX PRO 6000 Blackwell 3XS AI Workstation with RTX PRO 6000 Blackwell Max-Q NVIDIA DGX Station GB300
AI performance per GPU (FP4) 1,000 TOPS 1,824 TOPS 3,352 TOPS 4,000 TOPS 3,351 TOPS 20,000 TOPS
Memory per GPU 128GB 24GB 32GB 96GB 96GB 784GB
AI model size per GPU (FP4) 200 billion 38 billion 51 billion 153 billion 153 billion 1.2 trillion
GPU(s) 1 1 Up to 2 Up to 2 Up to 4 1
Maximum AI model size (FP4) 400 billion (2x DGX Spark) 38 billion 102 billion 307 billion 614 billion 1.2 trillion
Cost £ £ £ £££ £££ £££££

Conclusion

Thanks to its innovative SoC architecture, the DGX Spark is the only reasonably-priced desktop device that can develop AI models with a large number of parameters, such as generative AI and reasoning AI models. Being a desktop device, the DGX Spark is also a practical choice for organisations that are not able to upload their data into a cloud service. Having said that, if performance is more important to you than outright GPU memory, you should consider either an a (link to tab 2) or 3XS AI Development Workstation (link to tab 3).

Ready to buy?

Click the link below to view the range of AI development solutions. If you still have questions on how to select the perfect system, don't hesitate to contact one of our friendly advisors on 01204 474210 or at [email protected].

DGX Spark

3XS AI Development Laptop

Nvidia DGX Spark

3XS AI Development Laptops are custom-built for flexibility whilst maintaining the ability to effectively develop and debug your AI projects. Featuring an 18” QHD+ screen, an x86 CPU combined with the latest NVIDIA Blackwell GPU enabling you to identify the MVP (Minimum Viable Product). This mobile platform is ideal to learn valuable insights from failed models before rolling out your code to a compatible NVIDIA training system, such as an MGX, HGX or DGX server.

3XS Systems has been hand crafting PCs, workstations and servers for more than 30 years. We pioneered the AI dev box back in 2016, so we have a huge amount of experience in building highly reliable systems that deliver the most performance for your budget. Here are the key reasons to trust a 3XS AI Development Laptop.

NVIDIA Elite Partner

NVIDIA Elite Partner

Scan has been an accredited NVIDIA Elite Partner since 2017, awarded for our expertise in the areas of deep learning and AI.

AI Optimised

AI Optimised

Our in-house team includes data scientists who optimise the configuration and software stack of each system for AI workloads.

Trusted by you

Trusted by you

Scan 3XS Systems AI dev boxes are trusted by organisations including the NHS, University of Liverpool and University of Strathclyde.

7 Days Support

7 Days Support

Our technical support engineers are available seven days a week to help with any queries.

3 Years Warranty

3 Years Warranty

3XS Systems include a three-year warranty, so if anything goes faulty we’ll repair or replace it.

Architecture

The specifications feature a 24GB NVIDIA GeForce RTX 5090 GPU with an Intel Core Ultra 9 275HX CPU, plus up to 64GB of Corsair Vengeance 4,400MHz DDR5 system memory. An ultra-fast 2TB SSD is provided for the OS, caching and datasets.

GEFORCE RTX 5090
Architecture Blackwell
CUDA Cores 10,496
Tensor Cores 328
5th gen
RT Cores 82
4th gen
Memory 24GB
GDDR7
ECC Memory
Memory Controller 256-bit
NVLink
TDP 95W

The systems are soak tested with deep learning workloads and pre-installed with the latest Ubuntu operating system plus a custom software stack built on NVIDIA CUDA that includes Docker-CE, Nvidia-Docker2 and GPU-optimised libraries. Optional access to NVAIE for frameworks and applications is also available.

Relative Performance & Capability

A 3XS AI Development laptop has an FP4 performance of 1,824 TOPS which as you can see in the table below is higher than a DGX Spark, but with a much lower GPU memory capacity - 24GB compared to 128GB. However, these laptop devices are aimed at mobile AI development more than outright capacity or performance.

System NVIDIA DGX Spark 3XS AI Laptop with GeForce RTX 5090 3XS AI Workstation with GeForce RTX 5090 3XS AI Workstation with RTX PRO 6000 Blackwell 3XS AI Workstation with RTX PRO 6000 Blackwell Max-Q NVIDIA DGX Station GB300
AI performance per GPU (FP4) 1,000 TOPS 1,824 TOPS 3,352 TOPS 4,000 TOPS 3,351 TOPS 20,000 TOPS
Memory per GPU 128GB 24GB 32GB 96GB 96GB 784GB
AI model size per GPU (FP4) 200 billion 38 billion 51 billion 153 billion 153 billion 1.2 trillion
GPU(s) 1 1 Up to 2 Up to 2 Up to 4 1
Maximum AI model size (FP4) 400 billion (2x DGX Spark) 38 billion 102 billion 307 billion 614 billion 1.2 trillion
Cost £ £ £ £££ £££ £££££

Conclusion

Thanks to its portability, 3XS AI Development Laptops are the only solution that enables model development from anywhere - albeit with some limitations. It may be that you pair a laptop alongside a desktop device, such as a DGX Spark or 3XS AI Development Workstation, so work needn’t entirely stop when you are away from an office or lab environment.

Ready to buy?

Click the link below to view the range of AI development solutions. If you still have questions on how to select the perfect system, don't hesitate to contact one of our friendly advisors on 01204 474210 or at [email protected].

Pre-configured AI Development Laptops

3XS AI Development Workstation

3XS AI Development Workstation

3XS AI Development Workstations are custom-built to allow maximum tailoring and scaling for your models. Featuring multiple NVIDIA RTX PRO or GeForce RTX GPUs, they offer the ultimate in cost-effective AI model development, ready for rolling out your code to a compatible NVIDIA training system, such as an MGX, HGX or DGX server.

3XS Systems has been hand crafting PCs, workstations and servers for more than 30 years. We pioneered the AI dev box back in 2016, so we have a huge amount of experience in building highly reliable systems that deliver the most performance for your budget. Here are the key reasons to trust a 3XS AI Development Workstation.

NVIDIA Elite Partner

NVIDIA Elite Partner

Scan has been an accredited NVIDIA Elite Partner since 2017, awarded for our expertise in the areas of deep learning and AI.

AI Optimised

AI Optimised

Our in-house team includes data scientists who optimise the configuration and software stack of each system for AI workloads.

Whisper Quiet

Whisper Quiet

Only hear what matters – select configurations are watercooled and so are much quieter than air-cooled PCs.

Trusted by you

Trusted by you

Scan 3XS Systems AI dev boxes are trusted by organisations including the NHS, University of Liverpool and University of Strathclyde.

7 Days Support

7 Days Support

Our technical support engineers are available seven days a week to help with any queries.

3 Years Warranty

3 Years Warranty

3XS Systems include a three-year warranty, so if anything goes faulty we’ll repair or replace it.

Architecture

The flexible multi-GPU configuration supports either NVIDIA RTX PRO GPUs for maximum performance or consumer-grade NVIDIA GeForce GPUs for value for money. A simple comparison is detailed in the table below, but for further information please read our NVIDIA Workstation GPU Buyers Guide or our NVIDIA GeForce GPU Buyers Guide.

RTX PRO 6000 Blackwell RTX PRO 6000 Blackwell Max-Q RTX PRO 5000 Blackwell RTX PRO 4500 Blackwell RTX PRO 4000 Blackwell A800 RTX 5000 Ada RTX 4500 Ada RTX 4000 ADA GeForce RTX 5090 GeForce RTX 5080
Architecture Blackwell Blackwell Blackwell Blackwell Blackwell Ampere Ada Lovelace Ada Lovelace Ada Lovelace Blackwell Blackwell
CUDA Cores 24,064 24,064 12,800 10,496 8960 6,912 12,800 7,680 6,144 21,760 10,752
Tensor Cores 752
5th gen
752
5th gen
400
5th gen
328
5th gen
280
5th gen
432
3rd gen
400
4th gen
240
4th gen
192
4th gen
680
5th gen
336
5th gen
RT Cores 188
4th gen
188
4th gen
100
4th gen
82
4th gen
70
4th gen
0 100
3rd gen
60
3rd gen
48
3rd gen
170
4th gen
84
4th gen
Memory 96GB
GDDR7
96GB
GDDR7
48GB
GDDR7
32GB
GDDR7
24GB
GDDR7
40GB
HBM2
36GB
GDDR6
24GB
GDDR6
20GB
GDDR6
32GB
GDDR7
16GB
GDDR7
ECC Memory
Memory Controller 512-bit 512-bit 384-bit 256-bit 192-bit 5,120-bit 256-bit 192-bit 160-bit 512-bit 256-bit
NVLink 400GB/sec
TDP 600W 300W 300W 200W 140W 240W 250W 192W 130W 575W 360W

For maximum customisation, these GPUs can be combined with either AMD Ryzen 9 or Threadripper CPUs, or Intel Core Ultra 9 or Xeon-W CPUs. You can learn more about AMD and Intel CPUs by reading our AMD Desktop CPU Buyers Guide or our Intel Desktop CPU Buyers Guide. The systems support up to 512GB system memory and multiple SSDs. Networking is 10GbE as standard, with options for NVIDIA ConnectX SmartNICs at Ethernet or InfiniBand Speeds of up to 400Gb/s.

The systems are soak tested with deep learning workloads and pre-installed with the latest Ubuntu operating system plus a custom software stack built on NVIDIA CUDA that includes Docker-CE, Nvidia-Docker2 and GPU-optimised libraries. They are Proxmox-ready and feature optional access to NVAIE for frameworks and applications is also available.

Relative Performance & Capability

From the below table, 3XS AI Development Workstations have similar FP4 performance whether you select an RTX 5090 or either variety of RTX PRO 6000. However, the RTX 5090 is limited in model scale with only a third of the RTX PRO 6000’s memory. Furthermore, the RTX PRO 6000 Max-Q GPU has half the TDP of the regular RTX PRO 6000, allowing four cards to be installed in a single workstation providing capacity for much larger AI models - even outstripping the potential of a pair of DGX Spark devices.

System NVIDIA DGX Spark 3XS AI Laptop with GeForce RTX 5090 3XS AI Workstation with GeForce RTX 5090 3XS AI Workstation with RTX PRO 6000 Blackwell 3XS AI Workstation with RTX PRO 6000 Blackwell Max-Q NVIDIA DGX Station GB300
AI performance per GPU (FP4) 1,000 TOPS 1,824 TOPS 3,352 TOPS 4,000 TOPS 3,351 TOPS 20,000 TOPS
Memory per GPU 128GB 24GB 32GB 96GB 96GB 784GB
AI model size per GPU (FP4) 200 billion 38 billion 51 billion 153 billion 153 billion 1.2 trillion
GPU(s) 1 1 Up to 2 Up to 2 Up to 4 1
Maximum AI model size (FP4) 400 billion (2x DGX Spark) 38 billion 102 billion 307 billion 614 billion 1.2 trillion
Cost £ £ £ £££ £££ £££££

Conclusion

3XS AI Workstations are a supremely capable AI development solution, due to their flexibility and scalability. Starting with just a single GPU they can be scaled significantly and easily, as projects demand or budgets allow. The highest-spec configurations surpass even the DGX Spark’s GPU memory and couples it with class-leading performance - hence the much greater cost. This is the best solution if you need gradual scale, with performance and GPU memory capacity. If your models are guaranteed not to get too large or performance isn’t that important, then a DGX Spark should be considered, however if you have significant financial resource or an immediate need for the largest development platform, then only the NVIDIA DGX Station GB300 will suffice.

Ready to buy?

Click the links below to view the range of AI development solutions. If you still have questions on how to select the perfect system, don't hesitate to contact one of our friendly advisors on 01204 474210 or at [email protected].

Pre-configured AI Development Workstations Configure an AI Development Workstation
DGX Station GB300

The DGX Station GB300 is NVIDIA’s latest datacentre-at-the desktop scale appliance for developers looking for a solution to develop and fine-tune the largest generative, agentic or physical AI models. These are particularly challenging workloads as they have the largest memory footprint, so are too large to run on most existing GPUs. The DGX Station GB300 has been designed from the ground-up to provide a single extremely large memory pool. It achieves this by replacing the old discrete CPU and GPU paradigm with a SoC, combining CPU and GPU cores together in a single unit, linked to a coherent pool of memory.

Architecture

Known as the GB300 Grace Blackwell Superchip, this SoC comprises a Blackwell Ultra GPU and a Grace 72-core Neoverse V2 Arm CPU, sharing a coherent 288GB HBM3e GPU memory and 496GB LPDDR5X system memory, connected via a chip-to-chip link at 900GB/s. The GPU features 5th generation Tensor cores and multi-GPU (MIG) capability where up to seven discreetly containerised instances can be created from the single powerful GPU, making model development very scalable, flexible and secure. Standard networking includes an NVIDIA ConnectX-8 SuperNIC with Ethernet or InfiniBand speeds of up to 800Gb/s.

GB300
Architecture Blackwell Ultra
CUDA Cores TBC
Tensor Cores TBC
5th gen
RT Cores TBC
4th gen
Memory 496GB LPDDR5X + 288HBM3e
ECC Memory
Memory Controller TBC
C2C Link 900GB/s
TDP TBC

The DGX Station DGX GB300 runs DGX OS, which is a NVIDIA-customised version of Ubuntu Linux. This enables developers to unleash the full potential with a proven software platform built around a complete AI software stack including access to NVIDIA NIM microservices, Blueprints and AI Enterprise. This is of course the same software stack as other NVIDIA AI appliances, making it easy to transfer your newly-developed AI model onto more powerful MGX, HGX or DGX systems, whether they’re in your own datacentre or in the cloud.

Relative Performance & Capability

The DGX Station GB300 has an FP4 performance of 20,000 TOPS, which as you can see in the table below, far outperforms the other AI development hardware options. It also has the largest memory capacity and fastest networking to back up its suitability for the largest parameter AI models.

System NVIDIA DGX Spark 3XS AI Laptop with GeForce RTX 5090 3XS AI Workstation with GeForce RTX 5090 3XS AI Workstation with RTX PRO 6000 Blackwell 3XS AI Workstation with RTX PRO 6000 Blackwell Max-Q NVIDIA DGX Station GB300
AI performance per GPU (FP4) 1,000 TOPS 1,824 TOPS 3,352 TOPS 4,000 TOPS 3,351 TOPS 20,000 TOPS
Memory per GPU 128GB 24GB 32GB 96GB 96GB 784GB
AI model size per GPU (FP4) 200 billion 38 billion 51 billion 153 billion 153 billion 1.2 trillion
GPU(s) 1 1 Up to 2 Up to 2 Up to 4 1
Maximum AI model size (FP4) 400 billion (2x DGX Spark) 38 billion 102 billion 307 billion 614 billion 1.2 trillion
Cost £ £ £ £££ £££ £££££

Conclusion

Thanks to its ground-breaking SoC architecture, the DGX Station GB300 is the only desktop device that can develop AI models at enormous scale and great speed.

Ready to buy?

Click the links below to view the range of AI development solutions. If you still have questions on how to select the perfect system, don't hesitate to contact one of our friendly advisors on 01204 474210 or at [email protected].

DGX Station GB300

Proof of Concept

Any of these four solutions can be trialed free-of-charge in our proof-of-concept hosted environment. Your trial will involve secure access where you can use a sample of your own data for most realistic insights, and you’ll be guided by our expert data scientists to ensure you get the most from your PoC.

To arrange your PoC, contact our AI team.

Get Cloud Computing in 3 Simple Steps
Icon 1. Choose your GPU

Browse our available options or connect with a specialist to discuss a bespoke solution.

Icon 2. Rapid Provisioning

We’ll provision your environment and take you through a guided onboarding process.

Icon 3. Enjoy SCAN Cloud

You’re online! Set up, get to work, and access support anytime.

AI Development in the Cloud

Although this guide has been concerned with AI development hardware, we also supply all of these GPU-accelerated systems as virtual systems on our Scan Cloud platform.

Simple, Flexible Pricing with No Hidden Fees

No long-term commitments, no extra charges for storage or networking—experience the full benefits of the cloud without the drawbacks.

Reference Architecture for Unrivaled Performance

Harness the power of the latest GPUs for desktops or high-performance GPU clusters, from single GPUs to 8-way systems.

Networking and Storage Built for Performance

All GPU instances include NVMe storage and uncontended network ports, ensuring top-tier performance and data privacy.

Build It Your Way

Custom builds are SCAN’s specialty—every aspect of our Infrastructure-as-a-Service (IaaS) solutions is fully customizable, with UK-based solutions ensuring data sovereignty.

Browse the available Scan Cloud options on or contact our Cloud team.