
If you're reading this, there's a good chance your company has joined the growing list of organisations that have blocked access to ChatGPT, Claude, and other AI websites. You're not alone - according to a 2025 BlackBerry survey, 70% of companies now restrict access to generative AI tools, primarily to protect confidential information.
Here's what your IT department probably didn't tell you: you can run AI completely offline on your own laptop. No internet required. No data leaving your computer. No API calls to external servers. Just you, your machine, and a capable AI assistant that lives entirely on your hard drive.
Many corporate clients set up local AI solutions when cloud-based tools aren't an option - whether for compliance reasons, data sensitivity, or simply because IT said "no." This guide shows you exactly how to do the same thing, even if you've never touched a command line before.
The Reality Check
34.8% of employee ChatGPT inputs now contain sensitive data - up from just 11% in 2023. Your IT department isn't being paranoid; they're being responsible. Local AI is the solution that gives you productivity gains without the risk.
Before we dive into solutions, let's understand the problem. When you type a question into ChatGPT, here's what happens:
1. Data Leakage Is Not Theoretical
In 2025, security researchers discovered over 225,000 OpenAI and ChatGPT credentials for sale on dark web markets. These weren't hacked from OpenAI - they were harvested from employee devices using infostealer malware. Once attackers log in, they gain access to the complete chat history, exposing any sensitive business data previously shared.
Samsung learned this the hard way when employees uploaded proprietary semiconductor code to ChatGPT. Deutsche Bank blocked ChatGPT entirely while evaluating "how to best use these types of capabilities while ensuring the security of our and our client's data."
2. Compliance Requirements
Depending on your industry, using cloud AI may violate:
| Regulation | What It Covers | AI Risk |
|---|---|---|
| GDPR | EU personal data | Data leaving jurisdiction |
| HIPAA | Healthcare records | PHI exposure |
| SOC 2 | Security controls | Uncontrolled data flows |
| Privacy Act 1988 (AU) | Personal information | Overseas disclosure |
| Legal Privilege | Attorney-client comms | Waiver risk |
3. Intellectual Property Protection
Every prompt you type could potentially be used to train future AI models. Even if the provider says they won't, their terms of service can change. With local AI, this risk is zero - your data never leaves your machine.
Local AI means running artificial intelligence models directly on your computer, with no internet connection required. Think of it like this:
Key differences from cloud AI:
The trade-off? Local models are generally smaller and less capable than the massive models running on cloud servers with thousands of GPUs. But for everyday office tasks - email drafting, document summarisation, meeting notes, code assistance - they're more than capable.
After testing dozens of options, here are the four tools that actually work for corporate users:
| Metric | Ease of Use | Power | Improvement |
|---|---|---|---|
| LM Studio | Easiest (GUI) | High | Best for beginners |
| Ollama | Easy (CLI) | Highest | Best for power users |
| Jan.ai | Very Easy | Medium | Most polished UI |
| GPT4All | Easy | Medium | Best Windows experience |
What it is: A free desktop application that lets you download and run AI models with a ChatGPT-like interface. No coding required.
Why I recommend it first: In my experience setting up local AI for non-technical users, LM Studio has the lowest friction. You download it, double-click to install, search for a model, click download, and start chatting. That's it.
Key features:
Current version: 0.3.36 (as of January 2026)
Want a complete walkthrough? See our LM Studio Complete Beginner's Guide for step-by-step installation, interface tour, and troubleshooting tips.
What it is: A command-line tool that makes running local AI models as simple as Docker made running containers. One command to download, one command to run.
Why it matters: If you're comfortable with a terminal, Ollama is the most flexible and powerful option. It's also the foundation that many other tools build on.
Key features:
ollama run llama3.2)What it is: An open-source ChatGPT alternative that runs entirely offline. Jan focuses on privacy-first design with a beautiful, modern interface.
Why consider it: Jan was the easiest to install in my testing and has the most polished user interface. It's 100% free and open source under AGPL license.
Key features:
What it is: A desktop application from Nomic AI designed to run large language models on consumer hardware. It's specifically optimised for accessibility.
Why it's notable: GPT4All shines for non-technical users who want local AI. The UI is basic but functional, and models are typically 3-8GB, making them easy to download and run.
Key features:
Not all models are created equal. Here's what I recommend based on hundreds of client deployments:
Llama 3.2 8B (Meta)
Mistral 7B (Mistral AI)
Phi-4-mini (Microsoft)
Qwen 2.5 3B (Alibaba)
DeepSeek-R1-Distill-Qwen-32B
Llama 3.1 70B (Meta)
| Model | Size | RAM Needed | Best For |
|---|---|---|---|
| Phi-4-mini (3.8B) | ~2.5GB | 4-8GB | Lightweight tasks |
| Mistral 7B | ~4GB | 8GB | General use, fast |
| Llama 3.2 8B | ~5GB | 8-16GB | All-round best |
| DeepSeek-R1 7B | ~5GB | 8-16GB | Reasoning, coding |
| Llama 3.1 70B | ~40GB | 64GB+ | Maximum capability |
Here's the honest truth: local AI requires decent hardware. But "decent" doesn't mean "gaming PC with three GPUs."
To run smaller models (7B parameters or less):
| Metric | Model Class | Hardware Needed | Improvement |
|---|---|---|---|
| 3-7B parameters | Phi-4, Mistral 7B, Llama 8B | 8-16GB RAM, no GPU | Most laptops |
| 13-32B parameters | Llama 13B, DeepSeek 32B | 32GB RAM, RTX 3070+ | High-end laptops |
| 70B parameters | Llama 70B, DeepSeek 70B | 64GB+ RAM, RTX 4090 | Workstations |
Good news: Most business laptops from the past 3-4 years can run 7B models.
Requirements:
My recommendation: A Dell XPS, Lenovo ThinkPad, or HP EliteBook from 2022+ with 16GB RAM will handle Mistral 7B and Llama 3.2 8B comfortably.
Good news: Apple Silicon Macs are actually excellent for local AI due to unified memory architecture.
Requirements:
Performance note: An M4 Pro with 64GB RAM can run Qwen 2.5 32B at 11-12 tokens/second - that's production-ready speed.
If you have a Linux machine, you're probably technical enough to figure this out. But briefly:
If your laptop is more than 4-5 years old with 4-8GB RAM, you'll struggle. The number of useful AI models you can run locally on a 2019 laptop with 8GB RAM and no dedicated GPU is close to zero for practical purposes.
However, even a year-old 8-billion-parameter model is something you can get running on a reasonably modern notebook.
Let's get you running. I'll walk you through Ollama because it's the most versatile option, and once you understand it, other tools are easier.
Open Terminal (press Cmd + Space, type "Terminal", press Enter)
Install with Homebrew (if you have it):
brew install ollama
Or download directly from ollama.com and drag to Applications.
ollama --version
You should see something like ollama version 0.5.x
Download the installer from ollama.com
Run the installer - it's a standard "Next, Next, Finish" process
Open Command Prompt (press Windows key, type "cmd", press Enter)
Verify installation:
ollama --version
One command does everything:
curl -fsSL https://ollama.ai/install.sh | sh
Then verify:
ollama --version
Now the fun part. Let's download Llama 3.2 (a solid all-purpose model):
ollama pull llama3.2
This downloads the 8B parameter version (~5GB). Wait for it to complete - it might take 5-15 minutes depending on your internet speed.
ollama run llama3.2
You'll see a prompt like >>>. Type your question:
>>> Summarise this email in 3 bullet points: [paste your email text here]
To exit, type /bye or press Ctrl+D.
| Command | What It Does |
|---|---|
ollama list | Show downloaded models |
ollama run llama3.2 | Start chatting with a model |
ollama pull mistral | Download a new model |
ollama rm llama3.2 | Delete a model |
ollama serve | Start the API server |
If command lines aren't your thing, LM Studio is even easier.
Go to lmstudio.ai
Download for your operating system (Windows, macOS, or Linux)
Install:
Launch LM Studio
Click the magnifying glass (Discover tab) in the left sidebar
Search for "llama 3.2" or "mistral"
Click Download on the model you want
Wait for it to download (progress shows at the bottom)
Click the chat bubble icon in the left sidebar
Select your model from the dropdown at the top
Type your question in the chat box
Press Enter - you'll see the AI response stream in
That's it. You now have a private AI assistant on your laptop.
Here's what local AI is actually good for, based on my experience deploying these tools across accounting firms, legal practices, and corporate offices:
The prompt:
Write a professional email declining a meeting request. I'm too busy this week
but open to next week. Keep it brief and polite.
Works well because: Email is formulaic, and even 7B models handle it excellently.
The prompt:
Summarise this document in 5 key points:
[Paste your document text here]
Pro tip: For long documents, break them into chunks. Most local models have 8-32K token context limits (roughly 6,000-24,000 words).
The prompt:
Convert these rough meeting notes into a structured format with:
- Attendees
- Key decisions
- Action items with owners
- Next steps
Notes: [Paste your rough notes]
Why it works: Formatting and restructuring is a strength of local models.
The prompt:
Explain what this Excel formula does and suggest improvements:
=IF(AND(A1>100,B1<50),VLOOKUP(C1,Data!A:B,2,FALSE),"N/A")
Best models for code: DeepSeek-R1-Distill or CodeLlama variants.
The prompt:
I have a CSV with columns: Date, Product, Sales, Region.
Write a Python script to:
1. Calculate monthly sales by region
2. Find the top 3 products
3. Create a summary table
The prompt:
Translate this email to German, maintaining professional tone:
[Your email text]
Best models: Qwen 2.5 (supports 20+ languages) or Mistral.
The prompt:
Draft an executive summary for a quarterly report based on these points:
- Revenue up 12% YoY
- New client acquisitions: 47
- Churn rate decreased from 5% to 3.2%
- Major project delivered under budget
I'm not going to pretend local AI is as good as GPT-4 or Claude. It isn't. Here's the honest comparison:
| Metric | Local AI | Cloud AI (GPT-4/Claude) | Improvement |
|---|---|---|---|
| Privacy | 100% private | Data sent to servers | Local wins |
| Cost | Free after hardware | $20-100+/month | Local wins |
| Speed (7B model) | 5-15 tokens/sec | 50-100+ tokens/sec | Cloud wins |
| Capability | Good for routine tasks | Better reasoning/creativity | Cloud wins |
| Availability | Always available | Subject to outages | Local wins |
| Context length | 8-32K tokens typical | 128-200K tokens | Cloud wins |
Use local AI for:
Use cloud AI (when you can) for:
If you want to use local AI at work, here's how to approach it responsibly:
Don't say: "ChatGPT is blocked and I need it unblocked."
Do say: "I'd like to explore local AI tools that run entirely offline with no data leaving my device. Can we discuss whether tools like Ollama or LM Studio would meet our security requirements?"
Your organisation may still need:
When we help organisations implement local AI, we typically suggest:
Your company blocked ChatGPT for good reasons - data security matters. But that doesn't mean you should be left behind while AI transforms how work gets done.
Local AI gives you:
Yes, it's not as powerful as GPT-4. Yes, it requires a decent laptop. Yes, it takes 10-30 minutes to set up.
But once it's running, you have a private AI assistant that never shares your data, never costs extra, and never goes down because of server issues.
My recommendation: Start with LM Studio if you want the easiest experience, or Ollama if you're comfortable with command lines. Download Llama 3.2 8B or Mistral 7B. Try it for a week on non-sensitive tasks first. You'll be surprised how capable these local models have become.
Day 1: Install LM Studio or Ollama (10 minutes)
Day 2: Download Llama 3.2 8B and test with basic prompts (15 minutes)
Day 3: Try summarising a real document or drafting an email
Day 4: Experiment with different models for different tasks
Day 5: If it's working, talk to IT about formalising your use
Setting up local AI on one laptop is straightforward. Rolling it out across a team of 20, 50, or 200+ employees with proper governance, IT alignment, and compliance documentation is a different challenge entirely.
Solve8 helps Australian businesses implement private AI infrastructure that meets enterprise security requirements while keeping data within Australian borders.
What we offer:
| Metric | DIY Approach | With Solve8 | Improvement |
|---|---|---|---|
| Time to org-wide deployment | 2-4 months | 3-4 weeks | 4x faster |
| IT policy alignment | Research yourself | Templates provided | Hours saved |
| Model selection & testing | Trial and error | Expert guidance | Right fit first time |
| Staff training | Self-service | Included | Faster adoption |
Book a free 30-minute consultation →
No sales pitch. Just honest advice on whether local AI makes sense for your organisation.
Related Reading:
Sources:
Solve8 is an Australian AI consultancy helping businesses navigate the complex landscape of AI implementation. Based in Brisbane, serving clients across Australia. ABN: 84 615 983 732