Local AI security scanner. One command to find out if your Ollama, LM Studio, or self-hosted LLM setup is leaking like a sieve.
curl -fsSL https://iscooked.com/iscooked.com | bashOr download and run manually:
wget https://iscooked.com/iscooked.com
chmod +x iscooked.com
./iscooked.com| Check | What it looks for |
|---|---|
| Network Exposure | AI services listening on 0.0.0.0 instead of localhost |
| API Authentication | Ollama, LM Studio, Open WebUI running without auth |
| File Permissions | Model files and directories world-readable/writable |
| Docker Risks | AI containers running as root, privileged mode, host networking |
| GPU Exposure | NVIDIA/AMD device permissions and management endpoints |
| Telemetry | Active connections to known telemetry domains |
| Firewall Status | UFW, firewalld, iptables, nftables β is anything running? |
| SSL/TLS | AI services exposed over plain HTTP on non-localhost |
| Process Audit | AI processes and what user they're running as |
| Sensitive Files | .env files with API keys readable by other users |
| History & Logs | API keys leaked in shell history, world-readable log dirs |
| Ollama Config | OLLAMA_HOST, OLLAMA_ORIGINS, systemd service checks |
π₯ COOKED Ollama (port 11434) is listening on ALL interfaces
β
SAFE LM Studio (port 1234) is bound to localhost only
β WARMING UP Ollama API is responding without authentication
π₯ COOKED No active firewall detected!
π₯ COOKED Shell history contains ~3 potential API key(s)
YOUR COOKED SCORE
73% cooked [ββββββββββββββββββββββββββββββ ]
FULLY COOKED
3 critical 1 warnings 2 passed
You are absolutely cooked. Fix the critical issues above ASAP.
Your cooked score ranges from 0β100%:
- 0β14% β Looking Fresh. Your setup is locked down.
- 15β39% β Slightly Warm. A few things to tighten up.
- 40β69% β Medium Rare. Address those warnings.
- 70β100% β Fully Cooked. Fix the critical issues now.
- Bash 4+
- Standard Unix tools (ss/netstat, ps, stat, find)
- Optional:
curl(for API auth checks),docker(for container checks),nvidia-smi(for GPU checks) - Elevated privileges can improve some firewall and port checks
iscooked.com runs entirely on your machine. It makes no network requests, sends no telemetry, and phones home to absolutely nobody. The only network activity is checking if your local AI services are reachable on localhost.
PRs welcome! Some ideas:
- Add checks for more AI tools (KoboldCpp, TabbyAPI, Whisper, etc.)
- JSON output mode for CI/CD integration
- Auto-fix mode for common issues
- macOS-specific checks
- WSL-specific checks
MIT β do whatever you want with it.
Built by a cybersecurity engineer who runs local LLMs. iscooked.com