8 Commits

Author SHA1 Message Date
github-actions[bot]
7fd6cc9911 automated: Update flake.lock
Some checks failed
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 31s
Check Nix flake / Perform Nix flake checks (pull_request) Has been cancelled
Auto-generated by [update.yml][1] with the help of
[create-pull-request][2].

[1]: https://nayeonie.com/ahuston-0/nix-dotfiles/src/branch/main/.github/workflows/flake-update.yml
[2]: https://forgejo.stefka.eu/jiriks74/create-pull-request
2026-03-26 21:39:26 -04:00
623cad29a7 Merge pull request 're-add ollama to configuration' (#198) from feature/zed-models into main
All checks were successful
Check flake.lock / Check health of `flake.lock` (push) Successful in 9s
Check Nix flake / Perform Nix flake checks (push) Successful in 2m32s
Reviewed-on: #198
2026-03-26 13:02:30 -04:00
fde7963379 re-add ollama to configuration
All checks were successful
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 11s
Check Nix flake / Perform Nix flake checks (pull_request) Successful in 2m33s
2026-03-26 12:32:25 -04:00
e0f2f64886 Merge pull request 'ollama stuff' (#197) from feature/fwupd into main
All checks were successful
Check flake.lock / Check health of `flake.lock` (push) Successful in 8s
Check Nix flake / Perform Nix flake checks (push) Successful in 2m42s
Reviewed-on: #197
2026-03-26 12:27:26 -04:00
0036acbde3 devstral
All checks were successful
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 10s
Check Nix flake / Perform Nix flake checks (pull_request) Successful in 3m1s
2026-03-26 12:23:54 -04:00
ecdf223431 new models
All checks were successful
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 10s
Check Nix flake / Perform Nix flake checks (pull_request) Successful in 3m14s
2026-03-26 11:40:57 -04:00
6e6a8a205a remove gnome tools
All checks were successful
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 9s
Check Nix flake / Perform Nix flake checks (pull_request) Successful in 2m59s
2026-03-26 11:37:11 -04:00
342ff16158 ollama stuff
All checks were successful
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 7s
Check Nix flake / Perform Nix flake checks (pull_request) Successful in 3m16s
2026-03-26 11:32:04 -04:00
5 changed files with 56 additions and 53 deletions

View File

@@ -49,30 +49,5 @@
# Enable containerd for Kubernetes # Enable containerd for Kubernetes
virtualisation.containerd.enable = true; virtualisation.containerd.enable = true;
# Enable kubelet
#services.kubelet = {
# enable = true;
# extraFlags = {
# "pod-infra-container-image" = "registry.k8s.io/pause:3.9";
# };
#};
## Enable kubeadm for cluster initialization
#environment.etc."kubeadm.yaml".text = ''
# apiVersion: kubeadm.k8s.io/v1beta3
# kind: InitConfiguration
# localAPIEndpoint:
# advertiseAddress: 127.0.0.1
# bindPort: 6443
# ---
# apiVersion: kubeadm.k8s.io/v1beta3
# kind: ClusterConfiguration
# clusterName: ${config.services.kubernetes.clusterName}
# controlPlaneEndpoint: ${config.services.kubernetes.controlPlaneEndpoint}
# networking:
# serviceSubnet: 10.96.0.0/12
# podSubnet: 10.244.0.0/16
# dnsDomain: cluster.local
#'';
}; };
} }

View File

@@ -27,6 +27,7 @@
fd fd
file file
firefox firefox
# gestures replacement # gestures replacement
git git
glances glances

View File

@@ -26,10 +26,8 @@
programs.git.lfs.enable = false; programs.git.lfs.enable = false;
nixpkgs.config = { nixpkgs.config.packageOverrides = pkgs: {
packageOverrides = pkgs: { vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; };
vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; };
};
}; };
boot = { boot = {
@@ -100,13 +98,6 @@
smartd.enable = true; smartd.enable = true;
calibre-server.enable = false; calibre-server.enable = false;
# Kubernetes example configuration
# To enable Kubernetes, uncomment the following:
# kubernetes = {
# enable = true;
# clusterName = "palatine-hill-cluster";
# controlPlaneEndpoint = "localhost:6443";
# };
}; };
nix.gc.options = "--delete-older-than 150d"; nix.gc.options = "--delete-older-than 150d";

View File

@@ -15,27 +15,25 @@ in
"deepseek-r1:1.5b" "deepseek-r1:1.5b"
"deepseek-r1:32b" "deepseek-r1:32b"
"deepseek-r1:70b" "deepseek-r1:70b"
"qwen3" #"qwen3"
#"qwen3-coder-next" #"qwen3.5:latest"
"qwen3-coder-next" "qwen3-coder-next"
"lennyerik/zeta" "lennyerik/zeta"
"qwen2.5-coder:14b"
"qwen2.5-coder:32b"
"nomic-embed-text:latest" "nomic-embed-text:latest"
"llama4:scout"
"mistral:7b"
"minimax-m2.7:cloud"
"lfm2:24b" "lfm2:24b"
"glm-4.7-flash" "glm-4.7-flash"
"nemotron-cascade-2:30b"
"magistral"
"devstral-small-2"
]; ];
models = vars.primary_ollama; models = vars.primary_ollama;
environmentVariables = { environmentVariables = {
FLASH_ATTENTION = "1"; FLASH_ATTENTION = "1";
OLLAMA_KV_CACHE_TYPE = "q8_0"; OLLAMA_KV_CACHE_TYPE = "q8_0";
# Ollama memory configuration # Ollama memory configuration
OLLAMA_MAX_LOADED_MODELS = "3"; OLLAMA_MAX_LOADED_MODELS = "2";
OLLAMA_MAX_QUEUE = "512"; OLLAMA_MAX_QUEUE = "512";
OLLAMA_NUM_PARALLEL = "3"; OLLAMA_NUM_PARALLEL = "1";
# ROCm memory optimization # ROCm memory optimization
#HIP_VISIBLE_DEVICES = "0"; #HIP_VISIBLE_DEVICES = "0";

View File

@@ -63,8 +63,40 @@
"latex" "latex"
"terraform" "terraform"
"log" "log"
"context7-mcp-server"
"github-mcp-server"
]; ];
userSettings = { userSettings = {
context_servers = {
nixos = {
command = "nix";
args = [
"run"
"github:utensils/mcp-nixos"
"--"
];
};
};
language_models = {
ollama = {
api_url = "http://192.168.76.2:11434";
context_window = 128000;
# global keep alive doesnt work
#keep_alive = "15m";
available_models = [
{
name = "glm-4.7-flash";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "lfm2:24b";
max_tokens = 128000;
keep_alive = "15m";
}
];
};
};
colorize_brackets = true; colorize_brackets = true;
hard_tabs = false; hard_tabs = false;
vim_mode = true; vim_mode = true;
@@ -77,7 +109,7 @@
agent = { agent = {
default_model = { default_model = {
provider = "ollama"; provider = "ollama";
model = "qwen2.5-coder:latest"; model = "glm-4.7-flash";
}; };
favorite_models = [ ]; favorite_models = [ ];
model_parameters = [ ]; model_parameters = [ ];
@@ -89,13 +121,16 @@
journal = { journal = {
hour_format = "hour24"; hour_format = "hour24";
}; };
edit_preditions = { edit_predictions = {
provider = "open_ai_compatible_api"; provider = "ollama";
open_ai_compatible_api = { ollama = {
api_url = "http://localhost:11434/v1/completions"; #api_url = "http://192.168.76.2:11434/v1/completions";
model = "zeta:latest"; api_url = "http://192.168.76.2:11434";
prompt_format = "infer"; context_window = 128000;
model = "lennyerik/zeta";
prompt_format = "qwen";
max_requests = 64; max_requests = 64;
max_output_tokens = 256;
}; };
}; };
texlab = { texlab = {
@@ -144,6 +179,8 @@
# markdown # markdown
nodePackages.markdownlint-cli nodePackages.markdownlint-cli
# insert essential rust dependencies
# doom emacs dependencies # doom emacs dependencies
yaml-language-server yaml-language-server
nodePackages.typescript-language-server nodePackages.typescript-language-server
@@ -163,7 +200,7 @@
languagetool languagetool
# latex # latex
texlive.combined.scheme-medium texlive.combined.scheme-full
# dependencies for nix-dotfiles/hydra-check-action # dependencies for nix-dotfiles/hydra-check-action
nodejs_20 nodejs_20
@@ -184,5 +221,6 @@
# arch zed deps # arch zed deps
nixd nixd
uv
]; ];
} }