RunLocalModel.com

About RunLocalModel

By the RunLocalModel editorial team · Last updated: May 6, 2026

Short version: RunLocalModel.com is an independent, free tool that helps you find out whether your computer can run a particular open-source AI model locally - and how well - before you spend hours downloading anything.

Why we built this

Running large language models on your own machine has gone from "almost impossible" to "surprisingly practical" in just a few years. But the experience of getting started is still painful: you find a model on Hugging Face, see four different quantization options (Q4_K_M, Q5_K_M, Q6_K, Q8_0), six file sizes, and no clear answer to the only question that actually matters - "will this run on my computer, and how fast?"

Most people end up downloading an 8 GB file, watching their laptop crawl, and giving up. We wanted to fix the 30 seconds of decision-making that happens before the download starts. That is it. No newsletter, no paid tier, no account.

Who runs the site

RunLocalModel is maintained by a small, independent team of software engineers who have been running local LLMs since the early llama.cpp days. We work on the site in our spare time. We are not affiliated with NVIDIA, AMD, Apple, Hugging Face, Ollama, LM Studio, or any model vendor. When we recommend a model or a tool, it is because we have personally used it on the hardware we are talking about.

If you want to reach us - bug reports, hardware data corrections, partnership questions, anything - please use our contact page.

What the site actually does

Our editorial principles

  1. Show the math. Every estimate on the site is reproducible. The formulas are documented on the methodology page, not buried in our codebase.
  2. Be honest about uncertainty. VRAM estimation is not exact. Real performance depends on your OS, background apps, thermal throttling, the exact build of llama.cpp you are using, and even your GPU driver version. We try to flag this everywhere it matters.
  3. No invented benchmarks. When we cite a tokens-per-second number, it either comes from our own measurement on real hardware or from a clearly linked third-party source.
  4. No affiliate-driven recommendations. Our model and tool recommendations are based on what we actually use day to day. We do not change a recommendation because someone paid us to.
  5. Update or retract. If we get something wrong, we fix it and add a note to the changelog below.

Where our data comes from

What this site is not

Privacy

We do not require an account, do not collect personal information, and do not have a database of users. The site uses Google Analytics (anonymous, aggregated traffic only) and may serve ads via Google AdSense in the future. Full details are on our privacy page.

Changelog