How to beginner · 3 min read

How to use AI without internet

Quick answer
Use Ollama to run AI models locally on your machine without internet by installing the Ollama app and downloading models for offline use. You interact with these models via the ollama CLI or API, enabling AI-powered tasks without any network connection.

PREREQUISITES

  • macOS or Windows machine
  • Ollama app installed from https://ollama.com
  • Python 3.8+ for scripting with Ollama API
  • pip install requests

Setup Ollama locally

Download and install the Ollama app from https://ollama.com. This app allows you to run AI models entirely offline by downloading them to your local machine. After installation, open the app and download your preferred model (e.g., llama2 or mistral) for offline use.

bash
brew install ollama
output
==> Downloading https://ollama.com/download/ollama.pkg
==> Installing Ollama...
==> Ollama installed successfully

Step by step usage

Once you have Ollama installed and a model downloaded, you can run AI completions offline using the ollama CLI or via Python by calling the local API endpoint.

python
import requests

# Example: send a prompt to the local Ollama API
url = 'http://localhost:11434/completions'
headers = {'Content-Type': 'application/json'}
data = {
    'model': 'llama2',
    'prompt': 'Explain how to use AI without internet.',
    'max_tokens': 100
}

response = requests.post(url, json=data, headers=headers)
print(response.json()['completion'])
output
Explain how to use AI without internet by running models locally on your machine using Ollama. This allows you to perform AI tasks without any network connection.

Common variations

  • Use the ollama CLI directly: ollama run llama2 "Your prompt here"
  • Switch models by changing the model parameter in API calls.
  • Integrate Ollama with other languages by calling the local API endpoint http://localhost:11434/completions.
python
import subprocess

# Run prompt using CLI
result = subprocess.run(['ollama', 'run', 'llama2', 'What is AI?'], capture_output=True, text=True)
print(result.stdout)
output
AI is the simulation of human intelligence in machines that are programmed to think and learn.

Troubleshooting

  • If you see Connection refused, ensure the Ollama app is running and the local API server is active.
  • Check that the model is downloaded locally by running ollama list.
  • Restart the Ollama app if the API endpoint is unresponsive.

Key Takeaways

  • Ollama enables fully offline AI by running models locally on your machine.
  • Use the Ollama CLI or local API endpoint to interact with models without internet.
  • Ensure models are downloaded locally before offline use.
  • Troubleshoot connection issues by verifying Ollama app status and model availability.
Verified 2026-04 · llama2, mistral-large-latest
Verify ↗