ollama stuff #197
@@ -49,30 +49,5 @@
|
||||
# Enable containerd for Kubernetes
|
||||
virtualisation.containerd.enable = true;
|
||||
|
||||
# Enable kubelet
|
||||
#services.kubelet = {
|
||||
# enable = true;
|
||||
# extraFlags = {
|
||||
# "pod-infra-container-image" = "registry.k8s.io/pause:3.9";
|
||||
# };
|
||||
#};
|
||||
|
||||
## Enable kubeadm for cluster initialization
|
||||
#environment.etc."kubeadm.yaml".text = ''
|
||||
# apiVersion: kubeadm.k8s.io/v1beta3
|
||||
# kind: InitConfiguration
|
||||
# localAPIEndpoint:
|
||||
# advertiseAddress: 127.0.0.1
|
||||
# bindPort: 6443
|
||||
# ---
|
||||
# apiVersion: kubeadm.k8s.io/v1beta3
|
||||
# kind: ClusterConfiguration
|
||||
# clusterName: ${config.services.kubernetes.clusterName}
|
||||
# controlPlaneEndpoint: ${config.services.kubernetes.controlPlaneEndpoint}
|
||||
# networking:
|
||||
# serviceSubnet: 10.96.0.0/12
|
||||
# podSubnet: 10.244.0.0/16
|
||||
# dnsDomain: cluster.local
|
||||
#'';
|
||||
};
|
||||
}
|
||||
|
||||
@@ -27,6 +27,7 @@
|
||||
fd
|
||||
file
|
||||
firefox
|
||||
|
||||
# gestures replacement
|
||||
git
|
||||
glances
|
||||
|
||||
@@ -17,7 +17,6 @@
|
||||
./minio.nix
|
||||
./networking.nix
|
||||
./nextcloud.nix
|
||||
./ollama.nix
|
||||
#./plex
|
||||
./postgresql.nix
|
||||
./samba.nix
|
||||
@@ -26,10 +25,8 @@
|
||||
|
||||
programs.git.lfs.enable = false;
|
||||
|
||||
nixpkgs.config = {
|
||||
packageOverrides = pkgs: {
|
||||
vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; };
|
||||
};
|
||||
nixpkgs.config.packageOverrides = pkgs: {
|
||||
vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; };
|
||||
};
|
||||
|
||||
boot = {
|
||||
@@ -100,13 +97,6 @@
|
||||
smartd.enable = true;
|
||||
calibre-server.enable = false;
|
||||
|
||||
# Kubernetes example configuration
|
||||
# To enable Kubernetes, uncomment the following:
|
||||
# kubernetes = {
|
||||
# enable = true;
|
||||
# clusterName = "palatine-hill-cluster";
|
||||
# controlPlaneEndpoint = "localhost:6443";
|
||||
# };
|
||||
};
|
||||
|
||||
nix.gc.options = "--delete-older-than 150d";
|
||||
|
||||
@@ -15,27 +15,25 @@ in
|
||||
"deepseek-r1:1.5b"
|
||||
"deepseek-r1:32b"
|
||||
"deepseek-r1:70b"
|
||||
"qwen3"
|
||||
#"qwen3-coder-next"
|
||||
#"qwen3"
|
||||
#"qwen3.5:latest"
|
||||
"qwen3-coder-next"
|
||||
"lennyerik/zeta"
|
||||
"qwen2.5-coder:14b"
|
||||
"qwen2.5-coder:32b"
|
||||
"nomic-embed-text:latest"
|
||||
"llama4:scout"
|
||||
"mistral:7b"
|
||||
"minimax-m2.7:cloud"
|
||||
"lfm2:24b"
|
||||
"glm-4.7-flash"
|
||||
"nemotron-cascade-2:30b"
|
||||
"magistral"
|
||||
"devstral-small-2"
|
||||
];
|
||||
models = vars.primary_ollama;
|
||||
environmentVariables = {
|
||||
FLASH_ATTENTION = "1";
|
||||
OLLAMA_KV_CACHE_TYPE = "q8_0";
|
||||
# Ollama memory configuration
|
||||
OLLAMA_MAX_LOADED_MODELS = "3";
|
||||
OLLAMA_MAX_LOADED_MODELS = "2";
|
||||
OLLAMA_MAX_QUEUE = "512";
|
||||
OLLAMA_NUM_PARALLEL = "3";
|
||||
OLLAMA_NUM_PARALLEL = "1";
|
||||
|
||||
# ROCm memory optimization
|
||||
#HIP_VISIBLE_DEVICES = "0";
|
||||
|
||||
@@ -63,8 +63,40 @@
|
||||
"latex"
|
||||
"terraform"
|
||||
"log"
|
||||
"context7-mcp-server"
|
||||
"github-mcp-server"
|
||||
];
|
||||
userSettings = {
|
||||
context_servers = {
|
||||
nixos = {
|
||||
command = "nix";
|
||||
args = [
|
||||
"run"
|
||||
"github:utensils/mcp-nixos"
|
||||
"--"
|
||||
];
|
||||
};
|
||||
};
|
||||
language_models = {
|
||||
ollama = {
|
||||
api_url = "http://192.168.76.2:11434";
|
||||
context_window = 128000;
|
||||
# global keep alive doesnt work
|
||||
#keep_alive = "15m";
|
||||
available_models = [
|
||||
{
|
||||
name = "glm-4.7-flash";
|
||||
max_tokens = 128000;
|
||||
keep_alive = "15m";
|
||||
}
|
||||
{
|
||||
name = "lfm2:24b";
|
||||
max_tokens = 128000;
|
||||
keep_alive = "15m";
|
||||
}
|
||||
];
|
||||
};
|
||||
};
|
||||
colorize_brackets = true;
|
||||
hard_tabs = false;
|
||||
vim_mode = true;
|
||||
@@ -77,7 +109,7 @@
|
||||
agent = {
|
||||
default_model = {
|
||||
provider = "ollama";
|
||||
model = "qwen2.5-coder:latest";
|
||||
model = "glm-4.7-flash";
|
||||
};
|
||||
favorite_models = [ ];
|
||||
model_parameters = [ ];
|
||||
@@ -89,13 +121,16 @@
|
||||
journal = {
|
||||
hour_format = "hour24";
|
||||
};
|
||||
edit_preditions = {
|
||||
provider = "open_ai_compatible_api";
|
||||
open_ai_compatible_api = {
|
||||
api_url = "http://localhost:11434/v1/completions";
|
||||
model = "zeta:latest";
|
||||
prompt_format = "infer";
|
||||
edit_predictions = {
|
||||
provider = "ollama";
|
||||
ollama = {
|
||||
#api_url = "http://192.168.76.2:11434/v1/completions";
|
||||
api_url = "http://192.168.76.2:11434";
|
||||
context_window = 128000;
|
||||
model = "lennyerik/zeta";
|
||||
prompt_format = "qwen";
|
||||
max_requests = 64;
|
||||
max_output_tokens = 256;
|
||||
};
|
||||
};
|
||||
texlab = {
|
||||
@@ -144,6 +179,8 @@
|
||||
# markdown
|
||||
nodePackages.markdownlint-cli
|
||||
|
||||
# insert essential rust dependencies
|
||||
|
||||
# doom emacs dependencies
|
||||
yaml-language-server
|
||||
nodePackages.typescript-language-server
|
||||
@@ -163,7 +200,7 @@
|
||||
languagetool
|
||||
|
||||
# latex
|
||||
texlive.combined.scheme-medium
|
||||
texlive.combined.scheme-full
|
||||
|
||||
# dependencies for nix-dotfiles/hydra-check-action
|
||||
nodejs_20
|
||||
@@ -184,5 +221,6 @@
|
||||
|
||||
# arch zed deps
|
||||
nixd
|
||||
uv
|
||||
];
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user