
If you have read our complete guide to running AI offline at work, you know there are several tools for running AI locally: Ollama, Jan.ai, GPT4All, and LM Studio. Each has its place.
But if you have never touched a command line and just want something that works like ChatGPT—but runs entirely on your laptop—LM Studio is where you should start.
I have set up local AI for dozens of corporate clients. The pattern is always the same: non-technical users try Ollama, get stuck at the terminal, and give up. Then I show them LM Studio, and within 10 minutes they are chatting with a local AI model.
No coding. No command line. Just a clean interface, a download button, and a chat window.
What Makes LM Studio Different
LM Studio is not just a wrapper around command-line tools. It is a purpose-built desktop application with a proper user interface. Search for models, download them, and start chatting—all from one window. It is also completely free, with no hidden subscriptions or premium features.
This guide covers everything from first download to advanced features:
By the end, you will have a fully functional private AI assistant on your laptop that never sends data to the cloud.
Before downloading, let us make sure your hardware can handle local AI. Here is the honest truth about what you need.
| Metric | Minimum | Recommended | Improvement |
|---|---|---|---|
| RAM | 8GB (very limited) | 16GB or more | For 7-8B models |
| Storage | 10GB free | 50GB+ free | Models are 4-40GB each |
| CPU | Any modern CPU | M1+ or Intel 10th gen+ | AVX2 required on Intel |
| GPU | Not required | 4GB+ VRAM | Much faster with GPU |
If you have an Apple Silicon Mac (M1, M2, M3, or M4), you are in a great position. The unified memory architecture means your Mac can run surprisingly large models. An M1 MacBook Air with 16GB can comfortably run 7-8B parameter models.
Requirements:
Most business laptops from the past 3-4 years will work fine for smaller models.
Requirements:
LM Studio ships as an AppImage, so it works on most distributions without installation headaches.
Requirements:
This takes about 2 minutes.
Open your browser and navigate to lmstudio.ai. The site automatically detects your operating system and shows the appropriate download button.
Click the download button. The file is about 150-200MB depending on your platform:
macOS:
Windows:
Linux:
chmod +x LM-Studio-x.x.x.AppImage./LM-Studio-x.x.x.AppImageWhen you first open LM Studio, you will see a welcome screen. The app may check for updates—let it update if prompted. Version 0.3.36 (as of January 2026) includes important bug fixes and new model support.
LM Studio has a clean, modern interface. Once you understand where things are, everything clicks into place.
The sidebar contains five main tabs:
| Icon | Tab Name | What It Does |
|---|---|---|
| Home | Home | Welcome screen, recent models |
| Magnifying Glass | Discover | Search and download models |
| Chat Bubble | Chat | Your ChatGPT-like conversation interface |
| Server | Local Server | Run LM Studio as an API server |
| Folder | My Models | Manage downloaded models |
This is your "app store" for AI models. When you click the magnifying glass icon, you see a search bar at the top and curated model suggestions below.
What you will see:
This is where you spend most of your time. It looks and feels like ChatGPT:
Lists all models you have downloaded. From here you can:
This is where many beginners get stuck. There are thousands of models available, and picking the wrong one means either poor performance or failed loads.
If you have 16GB of RAM, download Llama 3.2 8B Instruct. Here is why:
Understanding Model Names:
Model filenames contain important information:
llama-3.2-8b-instruct-q4_k_m.gguf
│ │ │ │
│ │ │ └── Quantisation level (smaller file, slightly less accurate)
│ │ └── Fine-tuned for instruction following (conversations)
│ └── 8 billion parameters
└── Model family (Meta's Llama)
Quantisation quick guide:
Downloads can take 5-30 minutes depending on your internet speed and model size. The progress bar shows at the bottom of the screen. You can queue multiple downloads.
Once your model is downloaded, you are one click away from using it.
Click the chat bubble icon in the left sidebar. You will see an empty conversation area with a "Load a model to start chatting" prompt at the top.
What happens when loading:
Type your first message in the input box at the bottom and press Enter. Try something like:
Summarise the key benefits of working from home in 3 bullet points.
The AI will generate a response, streaming word by word. On a MacBook M1 with 16GB RAM, expect 10-20 tokens per second with an 8B model—fast enough to feel interactive.
Unlike single-prompt tools, LM Studio maintains conversation context. The AI remembers what you discussed earlier in the chat. Ask follow-up questions, request revisions, or change the topic.
Click the gear icon (or the collapsible panel on the right) to access model settings. These affect how the AI responds.
| Setting | What It Does | Recommended Value |
|---|---|---|
| Temperature | Controls randomness. Lower = more focused, higher = more creative | 0.7 for general use, 0.3 for factual tasks |
| Max Tokens | Maximum response length | 2048 for most tasks |
| Context Length | How much conversation history the model considers | 4096-8192 (hardware dependent) |
| GPU Layers | How much of the model runs on GPU | Auto, or increase for faster responses |
Temperature in practice:
The system prompt tells the AI how to behave. By default, most models use something like "You are a helpful assistant."
For work use, consider custom system prompts like:
You are a professional business assistant. Respond in a formal tone
suitable for corporate communication. Be concise and action-oriented.
Use Australian English spelling.
One of LM Studio's most useful features is document upload. You can attach files and ask the AI questions about them.
Example prompts after upload:
Summarise the key points of this document in 5 bullet points.
What are the payment terms mentioned in this contract?
List all the action items from these meeting notes.
LM Studio can run as a local API server, compatible with OpenAI's API format. This means any application designed for OpenAI can work with your local model instead.
http://localhost:1234/v1)Point applications to:
http://localhost:1234/v1Most OpenAI-compatible tools just need these two settings changed.
After helping many first-time users, these are the problems I see most often.
Cause: Not enough RAM for the selected model.
Fix:
Cause: Model running on CPU when GPU would be faster, or insufficient resources.
Fix:
Cause: Network issues or Hugging Face server problems.
Fix:
Cause: Incorrect chat template or model issue.
Fix:
If you are using LM Studio at work, here are the talking points and best practices for corporate environments.
When asking IT for approval, focus on the security story:
Good use cases:
Use cloud AI instead for:
You now have a working local AI assistant. Here are the next steps to get more value from it.
Once comfortable with Llama 3.2, try:
Save your favourite system prompts and settings as presets. Create different presets for:
If you use coding tools like VS Code with AI extensions, point them to LM Studio's local server for privacy-focused code assistance.
LM Studio removes the biggest barrier to local AI: complexity. You do not need to understand Python, Docker, or command-line tools. You just need a decent laptop and 10 minutes.
Is it as capable as GPT-4? No. But for routine work tasks—email drafting, document summarisation, meeting notes, quick questions—it is more than good enough. And your data never leaves your machine.
That trade-off is worth it for many professionals, especially those in regulated industries or companies that block cloud AI tools.
Download LM Studio, grab a Llama 3.2 model, and give it a try. The worst case is you learn something new. The best case is you gain a private AI assistant that costs nothing to use.
LM Studio is perfect for individual use, but rolling out local AI across a team of 10, 50, or 100+ employees requires proper planning. IT policy alignment, model governance, hardware requirements, and compliance documentation do not configure themselves.
Solve8 helps Australian businesses deploy private AI infrastructure that meets enterprise security requirements while keeping data within Australian borders.
What we offer:
| Metric | DIY Approach | With Solve8 | Improvement |
|---|---|---|---|
| Time to org-wide deployment | 2-4 months | 3-4 weeks | 4x faster |
| IT policy alignment | Research yourself | Templates provided | Hours saved |
| Model selection | Trial and error | Expert guidance | Right fit first time |
| Staff training | Self-service | Included | Faster adoption |
Book a free 30-minute consultation →
No sales pitch. Just honest advice on whether local AI makes sense for your organisation.
Related Resources:
Sources:
Solve8 is an Australian AI consultancy helping businesses navigate the complex landscape of AI implementation. Based in Brisbane, serving clients across Australia. ABN: 84 615 983 732