Skip to content

Bug: Ollama 4GB memory limit insufficient for 7B models #1958

@mrveiss

Description

@mrveiss

Problem

docker-compose.yml limits Ollama to 4GB RAM. The default model mistral:7b-instruct requires ~4.5-5.5GB (4GB weights + runtime overhead). This will cause OOM kills under any concurrent load.

Discovered During

Code review of PR #1930 (#1897)

Impact

Severity: high — default Docker deployment will OOM on inference

Fix

Increase to 8G minimum, or make configurable via env var. Dev override (PR #1935) should also relax this.

Metadata

Metadata

Assignees

No one assigned

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions