Can Your PC Run A.I. Ollama System Requirements Explained WG-AI-13 Ep. 2
So you wanna run Ollama?
Before you go downloading gigabytes of brain soup, you better make sure your machine can actually handle it.
In this episode, Izzy Rotgut lays out the real system requirements to run local LLMs using Ollama — from RAM and GPU to disk space and OS quirks.
We’re talkin CPUs, CUDA, VRAM, and goblin truth. No fluff. No hype. Just what works.
💻 In This Episode:
Minimum and recommended specs for Ollama
- Can you run it CPU-only? (Yes... slowly.)
- How much RAM do you actually need?
- What GPUs are supported (NVIDIA only!)
- Disk space requirements for popular models
- macOS, Windows, and Linux support
- What breaks — and what just cries loudly
📎 RESOURCES MENTIONED:
🔗 Official site: https://ollama.com
🔗 Model library: https://ollama.com/library
🔗 Install guide: https://ollama.com/download
🔗 Model size chart (Quantized): https://huggingface.co/collections/ollama/quantized-models-65d5bb226bc4661be12a6e58
🔗 Ollama docs: https://ollama.com/blog/ollama-cli
🔗 WSL for Windows: https://learn.microsoft.com/en-us/windows/wsl/install
⚠️ Coming Up Next:
A full walkthrough of the Ollama install process (Ep. 3)
A breakdown of which models run best on which hardware
And eventually? Izzy’s gonna earn herself a Quantum-Grade Phase-Brewed Blackhole Espresso Reactor — but only if you like, sub, and share.
🔖 Hashtags
#Ollama #LocalAI #RunAIOffline #IzzyRotgut #WGAI13
#AIPrivacy #LLMs #SystemRequirements #OpenSourceAI #GoblinEngineer
Видео Can Your PC Run A.I. Ollama System Requirements Explained WG-AI-13 Ep. 2 канала WG-AI-13 | Nerd News Across the Multiverse
Before you go downloading gigabytes of brain soup, you better make sure your machine can actually handle it.
In this episode, Izzy Rotgut lays out the real system requirements to run local LLMs using Ollama — from RAM and GPU to disk space and OS quirks.
We’re talkin CPUs, CUDA, VRAM, and goblin truth. No fluff. No hype. Just what works.
💻 In This Episode:
Minimum and recommended specs for Ollama
- Can you run it CPU-only? (Yes... slowly.)
- How much RAM do you actually need?
- What GPUs are supported (NVIDIA only!)
- Disk space requirements for popular models
- macOS, Windows, and Linux support
- What breaks — and what just cries loudly
📎 RESOURCES MENTIONED:
🔗 Official site: https://ollama.com
🔗 Model library: https://ollama.com/library
🔗 Install guide: https://ollama.com/download
🔗 Model size chart (Quantized): https://huggingface.co/collections/ollama/quantized-models-65d5bb226bc4661be12a6e58
🔗 Ollama docs: https://ollama.com/blog/ollama-cli
🔗 WSL for Windows: https://learn.microsoft.com/en-us/windows/wsl/install
⚠️ Coming Up Next:
A full walkthrough of the Ollama install process (Ep. 3)
A breakdown of which models run best on which hardware
And eventually? Izzy’s gonna earn herself a Quantum-Grade Phase-Brewed Blackhole Espresso Reactor — but only if you like, sub, and share.
🔖 Hashtags
#Ollama #LocalAI #RunAIOffline #IzzyRotgut #WGAI13
#AIPrivacy #LLMs #SystemRequirements #OpenSourceAI #GoblinEngineer
Видео Can Your PC Run A.I. Ollama System Requirements Explained WG-AI-13 Ep. 2 канала WG-AI-13 | Nerd News Across the Multiverse
Комментарии отсутствуют
Информация о видео
28 мая 2025 г. 5:00:19
00:04:55
Другие видео канала