Run Language Models on Your Computer with LM-Studio
A practical guide to running local models and picking the right one for speed or accuracy.
Good Morning,
You can do so much with AI. The building and DIY aspect also keeps getting more nuanced, and more powerful. Open-source AI is providing a new array of capabilities even at the local and individual level.
Iām a huge fan of Benjamin Marie and Iāve wanted to share more about his work for so long. Today, we finally have the chance. Ben is an independent AI researcher (LLM, NLP) with two really useful blogs and I have a huge respect for his work: (donāt let the funny names fool you, these are serious resources).
The Kaitchup ā AI on a Budget š
Hands on AI tutorials and news on how adapting language language models in a DIY setting to your tasks and hardware using the most recent techniques and models.
The Kaitchup publishes invaluable weekly tutorials with info thatās hard to find elsewhere.
By being a paid subscriber to The Kaitchup, you also get access to all the AI notebooks (160+), hands-on tutorials, and more in-depth analyses of recently published scientific papers.
Read The Salt š§
Reviews and in-depth analysis of bleeding edge AI research and how-tos. The Salt is a newsletter for readers who are curious about the Science behind AI. If you want to stay informed of recent progress in AI without reading much, The Salt is for you! I do my best to offer articles that might be interesting for a wide variety of readers.
Benjaminās technical and practical knowledge is invaluable depending on how deep down the rabbit-hole you want to go in DIY with models. Itās not overly technical but it is on technical topics, useful for a wide range of readers interested experimenting locally DIY with models or in small teams.
Selected Works
I asked him for a basic beginners tutorial on how to run LLMs locally (something I sometimes get questions about). Heās able to add so much practical know-how and insights into the latest models where for me he is an authority. If a new model comes out his opinion represents hands-on experience and being up to date on the latest scientific papers.
Qwen3-VL: DeepStack Fusion, Interleaved-MRoPE, and a Native 256K Interleaved Context Window.
Did the Model See the Benchmark During Training? Detecting LLM Contamination
Making LLMs Think Longer: Context, State, and Post-Training Tricks
Benjamin Marie is an independent researcher focused on hands-on AI and the tools around modern language models. He helps people and companies cut costs by adapting models to their specific tasks and hardware. I hope you learn something from it. While my work doesnāt touch on machine learning professionals that much, more and more individuals and small teals are playing with these open-source models locally.
So Iām very proud to be able to bring you a guide like this:
Run Language Models on Your Computer with LM-Studio
A practical guide to running local models and picking the right one for speed or accuracy.
Running large language models (LLMs) locally used to mean wrestling with the GPUās software layer (like CUDA), scattered model formats, and a lot of trial-and-error. Today, itās surprisingly approachable. With tools like Ollama or LM Studio, you can download a model, load it in a few clicks, and start chatting on your own machine, without sending prompts to a cloud service.
This article walks through the practical path from āinstalling the appā to ārunning my first local model,ā and then zooms out to the part that really matters: what determines whether a model runs smoothly (or not) on your hardware. Along the way, weāll cover installing LM Studio, the memory (simple) math behind model sizes, how to pick trustworthy GGUF builds and compression levels, how to sanity-check model output, and why āthinkingā models can be dramatically better on hard prompts while also being noticeably slower.
The goal is not to turn you into an engineer. Itās to give you enough intuition to choose models confidently, understand what an application like LM Studio is telling you, and avoid the most common āwhy is this slow / why is this wrongā surprises.








