blob: 758b22e742bff795a8cf8061ef3b9935803d5c7d (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
|
Ollama (Download and run large language models locally)
Ollama is an application which lets you run large language models
offline.
A list of models are available on ollama.com/library.
Optional dependencies like CUDA or ROCm will be automatically detected
during compilation of ollama libraries, if present.
CUDA=ON: building with CUDA, default is CUDA=OFF.
ROCM=ON: building with ROCm, default is ROCM=OFF.
Building ollama server and client requires network and
development/google-go-lang
|