Run DeepSeek R1 Locally with LM Studio – Complete Guide


Introduction
AI is evolving rapidly, and while cloud-based AI services are great, they come with privacy concerns, API limits, and server downtime. What if you could run DeepSeek R1 on your own machine without depending on external servers?
In this guide, I will walk you through installing LM Studio and running DeepSeek R1 1.5B model locally on Windows, macOS, and Linux. If you have ever wanted to experiment with AI models on your own machine, this guide is for you.
What is LM Studio?
LM Studio is an open-source tool that allows you to run large language models (LLMs) locally on your computer. It provides an easy interface to download, manage, and run AI models without needing advanced technical knowledge.
Why Use LM Studio?
No Internet Required – Run AI models offline on your machine
No API Limits – Use AI without any restrictions
Full Privacy – Your queries and data stay on your device
Supports Multiple Models – Run DeepSeek AI, Mistral, LLaMA, Falcon, and more
Installing LM Studio on Windows, macOS, and Linux
Installing LM Studio on Windows
Download LM Studio from the official website: LM Studio Download
Run the installer (.exe file) and follow the on-screen instructions
Once installed, open LM Studio
📌 Tip: If you are using a GPU, make sure your drivers are up to date for better performance.
Installing LM Studio on macOS
Open Terminal and install Homebrew (if not already installed):
/bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)"
Install LM Studio using Homebrew:
brew install lm-studio
Launch LM Studio from the Applications folder or Terminal:
lm-studio
📌 Tip: Mac users with Apple Silicon (M1/M2/M3) should use Metal acceleration for better AI model performance.
Installing LM Studio on Linux
Open Terminal and update your package manager:
sudo apt update && sudo apt upgrade -y
Download LM Studio from the official site or install using Flatpak:
flatpak install flathub ai.lmstudio.LMStudio
Run LM Studio:
flatpak run ai.lmstudio.LMStudio
📌 Tip: Some Linux distros may require additional CUDA dependencies for NVIDIA GPUs.
Downloading and Running DeepSeek R1 Model
Once LM Studio is installed, you need to download the DeepSeek R1 model and run it locally.
Download DeepSeek R1 Model
Open LM Studio
Go to the Model Catalog
Search for DeepSeek R1 1.5B
Click Download
📌 Tip: DeepSeek AI 1.5B model is lightweight compared to larger models like DeepSeek R1 7B or 67B, making it ideal for local experiments.
Running DeepSeek R1 Locally
Once the model is downloaded, follow these steps:
Open LM Studio
Navigate to the Local Models tab
Select DeepSeek R1 1.5B
Click Run Model
📌 Tip: If you have a powerful GPU, enable CUDA (for NVIDIA) or Metal (for Mac) for better performance.
Using DeepSeek AI for Local Queries
After the model is running, you can interact with DeepSeek AI directly:
Querying DeepSeek AI in LM Studio
Simply type your question in the chatbox and hit enter:
"What are the latest trends in AI development?"
DeepSeek AI will process your request locally without requiring an internet connection.
Using DeepSeek AI in Developer Mode (API Access)
If you want to integrate DeepSeek AI into your own applications, use LM Studio’s API mode:
Python Code to Use DeepSeek AI Locally
import requests
url = "http://localhost:8080/api/v1/chat"
data = {"prompt": "Generate Python code for a web scraper"}
response = requests.post(url, json=data)
print(response.json())
📌 Tip: You can deploy this locally for company-wide AI automation while keeping all data private.
Why Do We Need Local AI Models?
Running AI models locally has several advantages over cloud-based services like ChatGPT, DeepSeek, or OpenAI APIs:
Data Privacy – No information leaves your device
No Server Downtime – Unlike DeepSeek’s busy servers, local models are always available
Zero API Costs – AI inference is free on your own hardware
Custom Fine-Tuning – Modify AI models for specific tasks
However, there’s a trade-off: Larger AI models require massive hardware, making full-scale LLM deployment difficult on consumer devices.
📌 Example: DeepSeek AI’s 67B model requires a 480GB GPU, which is impractical for home use.
Final Thoughts & Next Steps
DeepSeek R1 1.5B model with LM Studio is a great way to experiment with AI locally without relying on external servers. While it may not match the performance of cloud-based LLMs, it provides greater control, privacy, and availability.
If you want to run larger AI models, consider upgrading your GPU, using cloud GPUs, or setting up AI clusters.
For more details watch this video
📌 What’s Next?
I’m working on setting up DeepSeek AI locally with a frontend AI agent—but due to hardware limitations (AMD ROCm compatibility issues), I faced some challenges.
Stay tuned for updates!
📌 Subscribe to my YouTube channel for more tutorials & insights: Tech With Asim YouTube
Would you like to see more tutorials on AI ? Let me know in the comments!
#DeepSeekAI #LMStudio #RunAIModelsLocally #TechWithAsim #SelfHostedAI #ArtificialIntelligence #AIForDevelopers #PrivateAI #NoCloudAI #OfflineAI
Subscribe to my newsletter
Read articles from Muhammad Asim directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by

Muhammad Asim
Muhammad Asim
With 13+ years of experience, I specialize in scalable web architectures, micro frontends, and high-performance applications. My journey began in 2012 as a graphics designer, creating 2D game assets during my internship, before transitioning into full-stack web development. Over the years, I've built SaaS applications, PWAs, fintech platforms, marketplaces, and e-commerce solutions. I’ve worked with startups, enterprises, and corporate giants like IBEX, Confiz, and Creative Chaos in technical and leadership roles. 💻 Highlights of Tech Stack & Expertise (There are alot more than the mentioned) ✔ Frontend: ReactJS, NextJS, TypeScript, AngularJS, Micro Frontends ✔ Backend: NodeJS, NestJS, Django, Laravel, Yii, CodeIgniter ✔ Databases: MariaDB, PostgreSQL, SQLite, Firebase, MongoDB ✔ DevOps & CI/CD: Docker, Jenkins, AWS (EC2, S3), Vagrant, CI/CD Pipelines ✔ 3rd Party Integrations: Stripe, PayPal, 2Checkout, Keycloak ✔ Architecture: Microservices, Scalable Web Apps, API Integrations 🚀 Career Highlights ✅ Transitioned from graphics design to full-stack development, mastering modern frameworks. ✅ Built and scaled SaaS, fintech, marketplaces, and e-commerce platforms. ✅ Led frontend teams, optimized architectures, and improved performance. ✅ Top-rated freelancer on Upwork. ✅ Managed tech communities, mentored developers, and conducted workshops. ✅ Co-founded two startups (Homemade Food Delivery & Online Artificial Jewelry) but paused due to work commitments. 🎯 Leadership & Mentorship I believe knowledge grows when shared, and after gaining extensive experience, it's time to give back to the tech community. ✔ I mentor junior developers, helping them transition into frontend and full-stack roles. ✔ Passionate about open-source contributions, tech blogging, and public speaking. ✔ Actively seeking mentorship programs and opportunities to conduct webinars. 🍲 Passion for Cooking & Travel Beyond tech, I love traveling and cooking! I enjoy experimenting with traditional & modern recipes and capturing my travel adventures. 📺 YouTube Channels: 👉 Dastarkhan Recipes – Culinary experiences & traditional dishes 👉 Muhammad Asim Vlogs – Documenting journeys across different places I believe in maintaining a balance between career growth and personal passions. 🤝 Let’s Connect! 🔹 Open to exciting projects, leadership roles, mentorship opportunities, and collaborations. 🔹 If you're into tech, startups, mentorship, or great food, let’s connect & chat! 🚀