install ollama
This commit is contained in:
parent
2e37cefe3e
commit
a93435fede
|
@ -14,6 +14,7 @@
|
|||
./services/amdgpu-clocks
|
||||
./services/flatpak
|
||||
./services/libinput
|
||||
./services/ollama
|
||||
./services/printers
|
||||
./services/udev
|
||||
./services/xserver
|
||||
|
|
|
@ -2,6 +2,17 @@
|
|||
# Get up and running with large language models locally.
|
||||
services.ollama = {
|
||||
enable = true;
|
||||
|
||||
# AMD GPU Support
|
||||
acceleration = "rocm";
|
||||
# 5700xt Support
|
||||
rocmOverrideGfx = "10.1.0";
|
||||
|
||||
# Language models to install
|
||||
loadModels = [
|
||||
"deepseek-coder-v2"
|
||||
"llama3"
|
||||
"mannix/llama3.1-8b-abliterated"
|
||||
];
|
||||
};
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue