install ollama
This commit is contained in:
parent
2e37cefe3e
commit
a93435fede
|
@ -14,6 +14,7 @@
|
||||||
./services/amdgpu-clocks
|
./services/amdgpu-clocks
|
||||||
./services/flatpak
|
./services/flatpak
|
||||||
./services/libinput
|
./services/libinput
|
||||||
|
./services/ollama
|
||||||
./services/printers
|
./services/printers
|
||||||
./services/udev
|
./services/udev
|
||||||
./services/xserver
|
./services/xserver
|
||||||
|
|
|
@ -2,6 +2,17 @@
|
||||||
# Get up and running with large language models locally.
|
# Get up and running with large language models locally.
|
||||||
services.ollama = {
|
services.ollama = {
|
||||||
enable = true;
|
enable = true;
|
||||||
|
|
||||||
|
# AMD GPU Support
|
||||||
acceleration = "rocm";
|
acceleration = "rocm";
|
||||||
|
# 5700xt Support
|
||||||
|
rocmOverrideGfx = "10.1.0";
|
||||||
|
|
||||||
|
# Language models to install
|
||||||
|
loadModels = [
|
||||||
|
"deepseek-coder-v2"
|
||||||
|
"llama3"
|
||||||
|
"mannix/llama3.1-8b-abliterated"
|
||||||
|
];
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue