Tool profile
Ollama
Ollama
A local runtime for downloading and serving large language models on personal or private infrastructure.
What it's used for
Ollama is used to run models locally for development, privacy-sensitive experiments, offline workflows, and simple self-hosted inference without a large cloud setup.
Categories
How you access it
See whether you access this through a vendor-hosted app, managed cloud, or official client.
LocalLow complexitySelf-hosted infra cost
Local runtime
Ollama is typically run on a local workstation or a privately controlled machine.
How you deploy or integrate it
See whether you can self-host it, deploy it in your stack, or integrate it through APIs and runtimes.
DockerMedium complexitySelf-hosted infra cost
Private servers, Docker, or VMs
Teams can host Ollama on a VM or container and expose it to internal apps that need private model access.