Ollama
Ollama
A local runtime for downloading and serving large language models on personal or private infrastructure.
Ollama is used to run models locally for development, privacy-sensitive experiments, offline workflows, and simple self-hosted inference without a large cloud setup.