Get up and running with Llama 2, Mistral, Gemma, and other large language models.

Edit Package ollama
https://ollama.com

Get up and running with Llama 2, Mistral, Gemma, and other large language models.

You can find a list of models available for use at https://ollama.com/library.

Source Files (show merged sources derived from linked package)
Filename Size Changed
_service 0000000226 226 Bytes
ollama-0.9.5.tar.gz 0008396484 8.01 MB
ollama-avoid-recomputing-special-vocabulary.patch 0000012001 11.7 KB
ollama-lib64-runner-path.patch 0000002494 2.44 KB
ollama-no-hip-redist.patch 0000001349 1.32 KB
ollama-user.conf 0000000159 159 Bytes
ollama.changes 0000056191 54.9 KB
ollama.service 0000000232 232 Bytes
ollama.spec 0000006016 5.88 KB
sysconfig.ollama 0000001410 1.38 KB
vendor.tar.zstd 0005615103 5.35 MB
Comments 1

Hzu's avatar

AMD users, install this one. This is the one you want with proper ROCm support. Thanks birdwatcher for taking the time to properly make ollama and ROCm modules fully available on Tumbleweed.

For Radeon 780M, there's no need to modify anything to get it running. However, due to some limitations imposed by ROCm and perhaps Ollama as well, you might be limited to 4096 MiB of VRAM. My GTT says I have 7000+ MiB of memory. However, ROCm only detects 4096 and will crash on most 7B models even though I have set UMA in BIOS to 16G.

As a workaround, you need to set a custom GTT size as well as TTM pool and page pool sizes to use your whole available VRAM. Instructions here: https://www.reddit.com/r/ROCm/comments/1g3lnuj/rocm_apu_680m_and_gtt_memory_on_arch/

There's an open PR in Ollama's repository as well: https://github.com/ollama/ollama/pull/6282

openSUSE Build Service is sponsored by