1 Commits

Author SHA1 Message Date
github-actions[bot]
bf4711e9a0 automated: Update flake.lock
All checks were successful
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 12s
Check Nix flake / Perform Nix flake checks (pull_request) Successful in 2m52s
Auto-generated by [update.yml][1] with the help of
[create-pull-request][2].

[1]: https://nayeonie.com/ahuston-0/nix-dotfiles/src/branch/main/.github/workflows/flake-update.yml
[2]: https://forgejo.stefka.eu/jiriks74/create-pull-request
2026-03-26 12:11:24 +00:00
5 changed files with 53 additions and 56 deletions

View File

@@ -49,5 +49,30 @@
# Enable containerd for Kubernetes # Enable containerd for Kubernetes
virtualisation.containerd.enable = true; virtualisation.containerd.enable = true;
# Enable kubelet
#services.kubelet = {
# enable = true;
# extraFlags = {
# "pod-infra-container-image" = "registry.k8s.io/pause:3.9";
# };
#};
## Enable kubeadm for cluster initialization
#environment.etc."kubeadm.yaml".text = ''
# apiVersion: kubeadm.k8s.io/v1beta3
# kind: InitConfiguration
# localAPIEndpoint:
# advertiseAddress: 127.0.0.1
# bindPort: 6443
# ---
# apiVersion: kubeadm.k8s.io/v1beta3
# kind: ClusterConfiguration
# clusterName: ${config.services.kubernetes.clusterName}
# controlPlaneEndpoint: ${config.services.kubernetes.controlPlaneEndpoint}
# networking:
# serviceSubnet: 10.96.0.0/12
# podSubnet: 10.244.0.0/16
# dnsDomain: cluster.local
#'';
}; };
} }

View File

@@ -27,7 +27,6 @@
fd fd
file file
firefox firefox
# gestures replacement # gestures replacement
git git
glances glances

View File

@@ -26,9 +26,11 @@
programs.git.lfs.enable = false; programs.git.lfs.enable = false;
nixpkgs.config.packageOverrides = pkgs: { nixpkgs.config = {
packageOverrides = pkgs: {
vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; }; vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; };
}; };
};
boot = { boot = {
loader.grub.device = "/dev/sda"; loader.grub.device = "/dev/sda";
@@ -98,6 +100,13 @@
smartd.enable = true; smartd.enable = true;
calibre-server.enable = false; calibre-server.enable = false;
# Kubernetes example configuration
# To enable Kubernetes, uncomment the following:
# kubernetes = {
# enable = true;
# clusterName = "palatine-hill-cluster";
# controlPlaneEndpoint = "localhost:6443";
# };
}; };
nix.gc.options = "--delete-older-than 150d"; nix.gc.options = "--delete-older-than 150d";

View File

@@ -15,25 +15,27 @@ in
"deepseek-r1:1.5b" "deepseek-r1:1.5b"
"deepseek-r1:32b" "deepseek-r1:32b"
"deepseek-r1:70b" "deepseek-r1:70b"
#"qwen3" "qwen3"
#"qwen3.5:latest" #"qwen3-coder-next"
"qwen3-coder-next" "qwen3-coder-next"
"lennyerik/zeta" "lennyerik/zeta"
"qwen2.5-coder:14b"
"qwen2.5-coder:32b"
"nomic-embed-text:latest" "nomic-embed-text:latest"
"llama4:scout"
"mistral:7b"
"minimax-m2.7:cloud"
"lfm2:24b" "lfm2:24b"
"glm-4.7-flash" "glm-4.7-flash"
"nemotron-cascade-2:30b"
"magistral"
"devstral-small-2"
]; ];
models = vars.primary_ollama; models = vars.primary_ollama;
environmentVariables = { environmentVariables = {
FLASH_ATTENTION = "1"; FLASH_ATTENTION = "1";
OLLAMA_KV_CACHE_TYPE = "q8_0"; OLLAMA_KV_CACHE_TYPE = "q8_0";
# Ollama memory configuration # Ollama memory configuration
OLLAMA_MAX_LOADED_MODELS = "2"; OLLAMA_MAX_LOADED_MODELS = "3";
OLLAMA_MAX_QUEUE = "512"; OLLAMA_MAX_QUEUE = "512";
OLLAMA_NUM_PARALLEL = "1"; OLLAMA_NUM_PARALLEL = "3";
# ROCm memory optimization # ROCm memory optimization
#HIP_VISIBLE_DEVICES = "0"; #HIP_VISIBLE_DEVICES = "0";

View File

@@ -63,40 +63,8 @@
"latex" "latex"
"terraform" "terraform"
"log" "log"
"context7-mcp-server"
"github-mcp-server"
]; ];
userSettings = { userSettings = {
context_servers = {
nixos = {
command = "nix";
args = [
"run"
"github:utensils/mcp-nixos"
"--"
];
};
};
language_models = {
ollama = {
api_url = "http://192.168.76.2:11434";
context_window = 128000;
# global keep alive doesnt work
#keep_alive = "15m";
available_models = [
{
name = "glm-4.7-flash";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "lfm2:24b";
max_tokens = 128000;
keep_alive = "15m";
}
];
};
};
colorize_brackets = true; colorize_brackets = true;
hard_tabs = false; hard_tabs = false;
vim_mode = true; vim_mode = true;
@@ -109,7 +77,7 @@
agent = { agent = {
default_model = { default_model = {
provider = "ollama"; provider = "ollama";
model = "glm-4.7-flash"; model = "qwen2.5-coder:latest";
}; };
favorite_models = [ ]; favorite_models = [ ];
model_parameters = [ ]; model_parameters = [ ];
@@ -121,16 +89,13 @@
journal = { journal = {
hour_format = "hour24"; hour_format = "hour24";
}; };
edit_predictions = { edit_preditions = {
provider = "ollama"; provider = "open_ai_compatible_api";
ollama = { open_ai_compatible_api = {
#api_url = "http://192.168.76.2:11434/v1/completions"; api_url = "http://localhost:11434/v1/completions";
api_url = "http://192.168.76.2:11434"; model = "zeta:latest";
context_window = 128000; prompt_format = "infer";
model = "lennyerik/zeta";
prompt_format = "qwen";
max_requests = 64; max_requests = 64;
max_output_tokens = 256;
}; };
}; };
texlab = { texlab = {
@@ -179,8 +144,6 @@
# markdown # markdown
nodePackages.markdownlint-cli nodePackages.markdownlint-cli
# insert essential rust dependencies
# doom emacs dependencies # doom emacs dependencies
yaml-language-server yaml-language-server
nodePackages.typescript-language-server nodePackages.typescript-language-server
@@ -200,7 +163,7 @@
languagetool languagetool
# latex # latex
texlive.combined.scheme-full texlive.combined.scheme-medium
# dependencies for nix-dotfiles/hydra-check-action # dependencies for nix-dotfiles/hydra-check-action
nodejs_20 nodejs_20
@@ -221,6 +184,5 @@
# arch zed deps # arch zed deps
nixd nixd
uv
]; ];
} }