update configuration of our ollama service
This commit is contained in:
parent
b0be0f9042
commit
2f545a818f
1 changed files with 25 additions and 1 deletions
|
|
@ -3,6 +3,10 @@
|
||||||
services.ollama = {
|
services.ollama = {
|
||||||
enable = true;
|
enable = true;
|
||||||
|
|
||||||
|
# User and group under which to run ollama
|
||||||
|
user = "ollama";
|
||||||
|
group = "ollama";
|
||||||
|
|
||||||
# AMD GPU Support
|
# AMD GPU Support
|
||||||
acceleration = "rocm";
|
acceleration = "rocm";
|
||||||
# 5700xt Support
|
# 5700xt Support
|
||||||
|
|
@ -11,9 +15,29 @@
|
||||||
# Language models to install
|
# Language models to install
|
||||||
loadModels = [
|
loadModels = [
|
||||||
"deepseek-r1:8b"
|
"deepseek-r1:8b"
|
||||||
"gemma3:12b"
|
"gemma3:4b"
|
||||||
"qwen3:8b"
|
"qwen3:8b"
|
||||||
"llama3:8b"
|
"llama3:8b"
|
||||||
|
|
||||||
|
# Coding models
|
||||||
|
"qwen2.5-coder:7b"
|
||||||
];
|
];
|
||||||
|
|
||||||
|
# Location to store models
|
||||||
|
models = "/srv/ollama/models";
|
||||||
|
};
|
||||||
|
|
||||||
|
# Enable the Open-WebUI server
|
||||||
|
services.open-webui = {
|
||||||
|
enable = true;
|
||||||
|
};
|
||||||
|
|
||||||
|
# Mount our subvolume for storage of models
|
||||||
|
fileSystems = {
|
||||||
|
"/srv/ollama" = {
|
||||||
|
device = "/dev/disk/by-label/data";
|
||||||
|
fsType = "btrfs";
|
||||||
|
options = ["subvol=srv-ollama" "compress=zstd"];
|
||||||
|
};
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue