feature/fwupd #196

Merged
ahuston-0 merged 3 commits from feature/fwupd into main 2026-03-23 23:24:37 -04:00

View File

@@ -12,19 +12,18 @@ in
package = pkgs.ollama; package = pkgs.ollama;
syncModels = true; syncModels = true;
loadModels = [ loadModels = [
"deepseek-r1:latest"
"deepseek-r1:1.5b" "deepseek-r1:1.5b"
"deepseek-r1:32b" "deepseek-r1:32b"
"deepseek-r1:70b"
"qwen3" "qwen3"
#"qwen3-coder-next" #"qwen3-coder-next"
"qwen3-coder" "qwen3-coder-next"
"lennyerik/zeta" "lennyerik/zeta"
"qwen2.5-coder:1.5b-base"
"qwen2.5-coder:14b" "qwen2.5-coder:14b"
"qwen2.5-coder:32b"
"nomic-embed-text:latest" "nomic-embed-text:latest"
"llama3.3:8b" "llama4:scout"
"mistral:7b" "mistral:7b"
"deepseek-v3.2:cloud"
"minimax-m2.7:cloud" "minimax-m2.7:cloud"
"lfm2:24b" "lfm2:24b"
"glm-4.7-flash" "glm-4.7-flash"
@@ -34,16 +33,16 @@ in
FLASH_ATTENTION = "1"; FLASH_ATTENTION = "1";
OLLAMA_KV_CACHE_TYPE = "q8_0"; OLLAMA_KV_CACHE_TYPE = "q8_0";
# Ollama memory configuration # Ollama memory configuration
OLLAMA_MAX_LOADED_MODELS = "2"; OLLAMA_MAX_LOADED_MODELS = "3";
OLLAMA_MAX_QUEUE = "4"; OLLAMA_MAX_QUEUE = "512";
OLLAMA_NUM_PARALLEL = "2"; OLLAMA_NUM_PARALLEL = "3";
# ROCm memory optimization # ROCm memory optimization
#HIP_VISIBLE_DEVICES = "0"; #HIP_VISIBLE_DEVICES = "0";
#ROCR_VISIBLE_DEVICES = "0"; #ROCR_VISIBLE_DEVICES = "0";
# context length for agents # context length for agents
OLLAMA_CONTEXT_LENGTH = "64000"; OLLAMA_CONTEXT_LENGTH = "128000";
}; };
openFirewall = true; openFirewall = true;
host = "0.0.0.0"; # don't want to make this available via load-balancer yet, so making it available on the local network host = "0.0.0.0"; # don't want to make this available via load-balancer yet, so making it available on the local network