Back to toolsOllama

Tool profile

Ollama

Ollama

A local runtime for downloading and serving large language models on personal or private infrastructure.

What it's used for

Ollama is used to run models locally for development, privacy-sensitive experiments, offline workflows, and simple self-hosted inference without a large cloud setup.

Categories

How you access it

See whether you access this through a vendor-hosted app, managed cloud, or official client.

LocalLow complexitySelf-hosted infra cost

Local runtime

Self-hosted

Ollama is typically run on a local workstation or a privately controlled machine.

How you deploy or integrate it

See whether you can self-host it, deploy it in your stack, or integrate it through APIs and runtimes.

DockerMedium complexitySelf-hosted infra cost

Private servers, Docker, or VMs

Self-hosted

Teams can host Ollama on a VM or container and expose it to internal apps that need private model access.