Tutorial14 min read

How to Build Your First AI Workstation (Step-by-Step Guide)

A complete walkthrough from parts list to running your first local LLM — hardware assembly, OS setup, NVIDIA drivers, CUDA, and Ollama configuration.

C

Compute Market Team

Our Top Pick

NVIDIA GeForce RTX 4090

$1,599 – $1,999

24GB GDDR6X | 16,384 | 1,008 GB/s

Buy on Amazon

What You'll Build

By the end of this guide, you'll have a fully working AI workstation running local LLMs, image generators, and AI agents — all on your own hardware with no cloud bills. No prior build experience required.

Note

This guide targets a mid-range build (~$3,000–$4,000 with an RTX 4090). You can swap components for any budget — see our AI workstation cost breakdown for alternative builds.

Step 1: Get Your Parts

Here's our recommended mid-range parts list:

ComponentOur PickWhy
GPUNVIDIA RTX 4090 24GBBest value for 24GB VRAM
CPUAMD Ryzen 7 7700X8 cores, great single-thread perf
MotherboardASUS TUF X670E-PLUSPCIe 5.0, dual M.2, solid VRMs
RAM64GB DDR5-5600 (2x32GB)Enough for datasets + system
StorageSamsung 990 Pro 4TBFast model loading, room for datasets
PSUCorsair RM1000x 1000W80+ Gold, fully modular, quiet
CaseFractal Design Meshify 2Excellent airflow for GPU heat
CoolerNoctua NH-D15Quiet, no pump failure risk

As Tim Dettmers writes in his widely cited GPU recommendations: "The most important single number for deep learning is the amount of GPU memory. It determines the maximum model size you can train and the batch sizes you can use." Tom's Hardware testing confirms the RTX 4090 remains the best consumer GPU for AI workloads in price-per-TFLOP when factoring in its 24GB VRAM ceiling.

Warning

The RTX 4090 is a massive card (3+ slot, 336mm long). Verify your case has clearance before buying. The Fractal Meshify 2 fits all current 4090 models.

Step 2: Assemble the Hardware

If you've never built a PC before, don't worry — it's straightforward. Think of it as adult LEGO with expensive pieces.

  1. Install CPU: Lift the motherboard retention arm, align the CPU triangle marker, drop it in (zero force), close the arm.
  2. Install RAM: Open the clips on the DIMM slots, align the notch, press firmly until both clips snap shut. Use slots A2 and B2 (2nd and 4th from CPU) for dual-channel.
  3. Install NVMe SSD: Unscrew the M.2 heatsink, insert the SSD at 30 degrees, push flat, screw in, replace heatsink.
  4. Mount motherboard: Install the I/O shield, align standoffs, screw in 9 screws (don't overtighten).
  5. Install PSU: Mount in the bottom compartment, route cables through the back. You'll need: 24-pin ATX, 8-pin CPU, and the 16-pin 12VHPWR cable for the GPU.
  6. Install GPU: Remove the necessary PCIe slot covers, insert the RTX 4090 into the top PCIe x16 slot, secure with screws, connect the 16-pin power cable.
  7. Install cooler: Apply thermal paste (pea-sized dot), mount the Noctua NH-D15, attach fans.
  8. Connect front panel: Power button, USB, audio headers — your motherboard manual has the pinout.

Step 3: Install the Operating System

You have two good options:

  • Ubuntu 24.04 LTS (recommended): Best ML framework support, native Docker, community tutorials. Most AI tools target Linux first.
  • Windows 11: Easier if you also want to use the machine for daily tasks. WSL2 provides a Linux environment for AI workloads.

Ubuntu Setup

  1. Download Ubuntu Desktop 24.04 LTS from ubuntu.com
  2. Flash to USB with Balena Etcher or Rufus
  3. Boot from USB, install with default settings
  4. After reboot, update the system:
sudo apt update && sudo apt upgrade -y
sudo reboot

Step 4: Install NVIDIA Drivers + CUDA

The GPU driver and CUDA toolkit are what let AI software talk to your hardware.

# Install NVIDIA driver
sudo apt install -y linux-headers-$(uname -r)
sudo apt install -y nvidia-driver-550

# Reboot to load the driver
sudo reboot

# Verify — you should see your RTX 4090 listed
nvidia-smi

You should see your GPU with driver version, temperature, and memory info. If it works, install CUDA:

# Install CUDA Toolkit
sudo apt install -y nvidia-cuda-toolkit

# Verify CUDA
nvcc --version

Warning

If nvidia-smi fails: disable Secure Boot in BIOS, reboot, and try again. Secure Boot blocks unsigned kernel modules including NVIDIA drivers.

Step 5: Install Ollama and Run Your First Model

Ollama is the fastest way to start running LLMs locally. One command to install, one command to run.

# Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh

# Run Llama 3.1 8B (downloads ~4.7GB on first run)
ollama run llama3.1

# Try a larger model if you have the VRAM
ollama run llama3.1:70b

That's it — you're now running AI locally. The 8B model works on any GPU with 6GB+ VRAM. The 70B model needs 24GB+ (your RTX 4090 handles it perfectly in 4-bit quantization). According to Puget Systems' workstation benchmarks, a properly configured RTX 4090 build achieves over 120 tokens/sec on Llama 3 8B — fast enough for real-time conversational AI.

Step 6: Install More AI Tools

Stable Diffusion (Image Generation)

# Install ComfyUI (recommended Stable Diffusion interface)
git clone https://github.com/comfyanonymous/ComfyUI.git
cd ComfyUI
pip install -r requirements.txt
python main.py

Open http://localhost:8188 in your browser. Download SDXL or Flux models from HuggingFace and start generating images.

Open WebUI (ChatGPT-like Interface for Ollama)

# Run Open WebUI via Docker
docker run -d -p 3000:8080 \
  --add-host=host.docker.internal:host-gateway \
  -v open-webui:/app/backend/data \
  --name open-webui \
  ghcr.io/open-webui/open-webui:main

Open http://localhost:3000 — you now have a ChatGPT-like interface running entirely on your hardware.

Step 7: Optimize Performance

  • Monitor GPU: Run watch -n 1 nvidia-smi to track utilization, temperature, and VRAM usage
  • Set GPU power limit: If thermals are an issue, sudo nvidia-smi -pl 350 reduces heat with minimal performance loss
  • Use quantized models: 4-bit (Q4_K_M) models use ~4x less VRAM with only 5–10% quality loss
  • Enable persistence mode: sudo nvidia-smi -pm 1 keeps the GPU initialized for faster cold starts

You're Done — What's Next?

You now have a complete AI workstation. From here:

tutorialbuildworkstationsetupbeginnerNVIDIAOllama

More from the blog

Stay ahead in AI hardware

Weekly deals, GPU reviews, and build guides. No spam.

Unsubscribe anytime. We respect your inbox.