Skip to content

We're upgrading our operations to serve you better. Orders ship as usual from Laval, QC. Questions? Contact us

Bitcoin accepted at checkout  |  Ships from Laval, QC, Canada  |  Expert support since 2016

Category: AI Models

Coverage of open-source AI models: Llama, Gemma, Mistral, Qwen, DeepSeek, Stable Diffusion, FLUX, Whisper.

AI

Used RTX 3090 for LLMs in 2026: Still King?

24 GB of VRAM at $600–$800 used. For LLMs under 70B parameters at Q4–Q5 quants, the RTX 3090 is still the pleb standard in 2026. Here’s the head-to-head vs 4090, 5090, P40, and A5000, plus a buying checklist.

AI

The Pleb’s Guide to Self-Hosted AI

Self-hosted AI isn’t as easy as opening ChatGPT — but for plebs who already run nodes and miners, the learning curve is half what it looks like. Here’s the whole picture before you install anything.

AI

LM Studio vs Ollama vs llama.cpp: Which Runner for Plebs?

Three excellent open-source runners. Three different plebs. llama.cpp is the foundation Gerganov built. Ollama wraps it for daemon simplicity. LM Studio wraps it in a polished GUI. Here’s the 15-minute decision guide.

AI

Open WebUI: The ChatGPT Experience, But Yours

The terminal is fine for testing, unusable for daily driving. Open WebUI is the ChatGPT-style interface that plugs into your local Ollama — multi-user, RAG, web search, reachable from anywhere over Tailscale. One Docker command; your Hashcenter becomes your private ChatGPT.

AI

ComfyUI for Plebs: Your First Local Image Generation

You installed Ollama and got local chat. Time for local image generation. ComfyUI runs SDXL, SD 3.5, and FLUX.1 on hardware you already own — the Midjourney/DALL-E subscription you can cancel. Here’s the pleb on-ramp.

AI

GGUF, Q4, Q8, fp16: A Pleb’s Guide to LLM Quantization

Quantization is lossy compression for LLMs — same idea as JPEG for photos. It’s the reason a used 3090 runs 70B models and an 8 GB laptop runs Phi-3.5. Here’s what the Q4_K_M and GGUF suffixes actually mean, and which quant to pick for your rig.