Local LLM Configuration Auditor
You downloaded Ollama, pulled a model, and started chatting. But did you check who else on your network can too? This free, open-source scanner finds every local AI endpoint on your machine or network and audits its security — before someone else finds it first.
The Problem
The local AI boom is real. Ollama has millions of downloads. LM Studio runs on every developer's laptop. vLLM powers production inference. But security hasn't kept up:
-
Ollama binds to
0.0.0.0:11434with no authentication — anyone can pull models, read your prompts, and run inference on your GPU - LM Studio exposes an OpenAI-compatible API with no access controls
- Most platforms leak their version, list all loaded models, and expose system prompts through their APIs
This tool scans for all of that in seconds and generates a detailed report with fix recommendations.
Quick Start
Three commands and you're scanning. No signup, no API keys, no cloud dependencies.
On Windows, you can also double-click start.bat
Scan Modes
Choose how you want to scan when you run the tool:
reports/
7 Security Checks Per Service
Each discovered LLM endpoint is audited against these checks:
| # | Check | Severity | Why It Matters |
|---|---|---|---|
| 1 | Unauthenticated API Access | Critical | Anyone can use your LLM without credentials |
| 2 | Unauthorized Model Pull | Critical | Attackers can download arbitrary models via Ollama's /api/pull |
| 3 | Model Enumeration | High | All loaded models are visible to anyone |
| 4 | System Prompt Extraction | High | Your model's system prompts and configs are exposed |
| 5 | Non-Localhost Binding | High | Service is reachable from outside your machine |
| 6 | CORS Misconfiguration | Medium | Access-Control-Allow-Origin: * enables browser-based attacks |
| 7 | Version Disclosure | Low | Software version exposed, useful for targeted exploits |
Deep Local Machine Scan
When scanning localhost, the tool goes beyond port scanning with 5 detection methods — it even finds LLMs that are installed but not currently running:
Supported Platforms
Ollama :11434
Most popular local LLM runner. Default config exposes API to all interfaces with zero auth.
LM Studio :1234
Desktop app with OpenAI-compatible server. No built-in access controls.
vLLM :8000
High-performance inference engine. Often deployed in production without auth.
LocalAI :8080
Drop-in OpenAI replacement. Supports multiple model backends.
llama.cpp :8080
Lightweight C++ inference server. Minimalist, often without security features.
text-generation-webui :7860
Gradio-based UI with API. Multiple exposed interfaces and ports.
Features
python run.py and pick a target
requests + optionally psutil
logs/ for review
Authorized Use Only
This tool is for authorized security auditing only. Only scan networks and systems you own or have explicit written permission to test. Unauthorized scanning may violate local laws. The authors are not responsible for misuse.
Frequently Asked Questions
requests library (installed via pip). For full deep-scan capabilities (process scanning, port harvesting), psutil is recommended but optional — the tool gracefully degrades without it. Works on Windows, macOS, and Linux.
Need Help Securing Your AI Infrastructure?
Our team can help you lock down local LLM deployments, implement authentication, configure network isolation, and audit your AI supply chain.
Book a Free Consultation