{"id":6933,"date":"2025-07-19T22:25:10","date_gmt":"2025-07-19T22:25:10","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/6933\/"},"modified":"2025-07-19T22:25:10","modified_gmt":"2025-07-19T22:25:10","slug":"how-to-run-an-llm-on-your-laptop","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/6933\/","title":{"rendered":"How to run an LLM on your laptop"},"content":{"rendered":"\n<p>For Pistilli, opting for local models as opposed to online chatbots has implications beyond privacy. \u201cTechnology means power,\u201d she says. \u201cAnd so who[ever] owns the technology also owns the power.\u201d States, organizations, and even individuals might be motivated to disrupt the concentration of AI power in the hands of just a few companies by running their own local models.<\/p>\n<p>Breaking away from the big AI companies also means having more control over your LLM experience. Online LLMs are constantly shifting under users\u2019 feet: Back in April, ChatGPT <a href=\"https:\/\/openai.com\/index\/sycophancy-in-gpt-4o\/\" rel=\"nofollow noopener\" target=\"_blank\">suddenly started sucking up<\/a> to users far more than it had previously, and just last week Grok started calling itself MechaHitler on X.<\/p>\n<p>Providers tweak their models with little warning, and while those tweaks might sometimes improve model performance, they can also cause undesirable behaviors. Local LLMs may have their quirks, but at least they are consistent. The only person who can change your local model is you.<\/p>\n<p>Of course, any model that can fit on a personal computer is going to be less powerful than the premier online offerings from the major AI companies. But there\u2019s a benefit to working with weaker models\u2014they can inoculate you against the more pernicious limitations of their larger peers. Small models may, for example, hallucinate more frequently and more obviously than Claude, GPT, and Gemini, and seeing those hallucinations can help you build up an awareness of how and when the larger models might also lie.<\/p>\n<p>\u201cRunning local models is actually a really good exercise for developing that broader intuition for what these things can do,\u201d Willison says.<\/p>\n<p>  How to get started  <\/p>\n<p>Local LLMs aren\u2019t just for proficient coders. If you\u2019re comfortable using your computer\u2019s command-line interface, which allows you to browse files and run apps using text prompts, <a href=\"https:\/\/ollama.com\/\" rel=\"nofollow noopener\" target=\"_blank\">Ollama<\/a> is a great option. Once you\u2019ve installed the software, you can download and run any of the hundreds of models they offer with a <a href=\"https:\/\/github.com\/ollama\/ollama\/blob\/main\/README.md#quickstart\" rel=\"nofollow noopener\" target=\"_blank\">single command<\/a>.\u00a0<\/p>\n<p>If you don\u2019t want to touch anything that even looks like code, you might opt for <a href=\"https:\/\/lmstudio.ai\/\" rel=\"nofollow noopener\" target=\"_blank\">LM Studio<\/a>, a user-friendly app that takes a lot of the guesswork out of running local LLMs. You can browse models from Hugging Face from right within the app, which provides plenty of information to help you make the right choice. Some popular and widely used models are tagged as \u201cStaff Picks,\u201d and every model is labeled according to whether it can be run entirely on your machine\u2019s speedy GPU, needs to be shared between your GPU and slower CPU, or is too big to fit onto your device at all. Once you\u2019ve chosen a model, you can download it, load it up, and start interacting with it using the app\u2019s chat interface.<\/p>\n<p>As you experiment with different models, you\u2019ll start to get a feel for what your machine can handle. According to Willison, every billion model parameters require about one GB of RAM to run, and I found that approximation to be accurate: My own 16 GB laptop managed to run Alibaba\u2019s <a href=\"https:\/\/huggingface.co\/Qwen\/Qwen3-14B\" rel=\"nofollow noopener\" target=\"_blank\">Qwen3 14B<\/a> as long as I quit almost every other app. If you run into issues with speed or usability, you can always go smaller\u2014I got reasonable responses from Qwen3 8B as well.<\/p>\n","protected":false},"excerpt":{"rendered":"For Pistilli, opting for local models as opposed to online chatbots has implications beyond privacy. \u201cTechnology means power,\u201d&hellip;\n","protected":false},"author":2,"featured_media":6934,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-6933","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/6933","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=6933"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/6933\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/6934"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=6933"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=6933"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=6933"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}