1 Commits

Author SHA1 Message Date
github-actions[bot]
bf4711e9a0 automated: Update flake.lock
All checks were successful
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 12s
Check Nix flake / Perform Nix flake checks (pull_request) Successful in 2m52s
Auto-generated by [update.yml][1] with the help of
[create-pull-request][2].

[1]: https://nayeonie.com/ahuston-0/nix-dotfiles/src/branch/main/.github/workflows/flake-update.yml
[2]: https://forgejo.stefka.eu/jiriks74/create-pull-request
2026-03-26 12:11:24 +00:00
7 changed files with 36 additions and 115 deletions

View File

@@ -46,10 +46,10 @@
kubernetes kubernetes
]; ];
## Enable containerd for Kubernetes # Enable containerd for Kubernetes
#virtualisation.containerd.enable = true; virtualisation.containerd.enable = true;
## Enable kubelet # Enable kubelet
#services.kubelet = { #services.kubelet = {
# enable = true; # enable = true;
# extraFlags = { # extraFlags = {

View File

@@ -1,7 +1,6 @@
{ {
lib, lib,
pkgs, pkgs,
config,
... ...
}: }:
{ {
@@ -18,7 +17,6 @@
./stylix.nix ./stylix.nix
./wifi.nix ./wifi.nix
./zerotier.nix ./zerotier.nix
../palatine-hill/ollama.nix
]; ];
time.timeZone = "America/New_York"; time.timeZone = "America/New_York";
@@ -41,19 +39,6 @@
sops.age.sshKeyPaths = [ "/etc/ssh/ssh_host_ed25519_key" ]; sops.age.sshKeyPaths = [ "/etc/ssh/ssh_host_ed25519_key" ];
services = { services = {
ollama = {
package = lib.mkForce pkgs.ollama-rocm;
models = lib.mkForce "${config.services.ollama.home}/models";
loadModels = lib.mkForce [
"deepseek-r1:1.5b"
"lennyerik/zeta"
"nomic-embed-text:latest"
"glm-4.7-flash"
"magistral"
"devstral-small-2"
"starcoder2:7b"
];
};
flatpak.enable = true; flatpak.enable = true;
calibre-web = { calibre-web = {
# temp disable this # temp disable this

View File

@@ -27,7 +27,6 @@
fd fd
file file
firefox firefox
# gestures replacement # gestures replacement
git git
glances glances
@@ -35,8 +34,12 @@
grim grim
htop htop
hwloc hwloc
ipmiview
iperf3 iperf3
# ipscan # ipscan
javaPackages.compiler.temurin-bin.jdk-25
javaPackages.compiler.temurin-bin.jdk-21
javaPackages.compiler.temurin-bin.jdk-17
jp2a jp2a
jq jq
kdePackages.kdenlive kdePackages.kdenlive

View File

@@ -17,6 +17,7 @@
./minio.nix ./minio.nix
./networking.nix ./networking.nix
./nextcloud.nix ./nextcloud.nix
./ollama.nix
#./plex #./plex
./postgresql.nix ./postgresql.nix
./samba.nix ./samba.nix
@@ -25,9 +26,11 @@
programs.git.lfs.enable = false; programs.git.lfs.enable = false;
nixpkgs.config.packageOverrides = pkgs: { nixpkgs.config = {
packageOverrides = pkgs: {
vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; }; vaapiIntel = pkgs.vaapiIntel.override { enableHybridCodec = true; };
}; };
};
boot = { boot = {
loader.grub.device = "/dev/sda"; loader.grub.device = "/dev/sda";
@@ -97,6 +100,13 @@
smartd.enable = true; smartd.enable = true;
calibre-server.enable = false; calibre-server.enable = false;
# Kubernetes example configuration
# To enable Kubernetes, uncomment the following:
# kubernetes = {
# enable = true;
# clusterName = "palatine-hill-cluster";
# controlPlaneEndpoint = "localhost:6443";
# };
}; };
nix.gc.options = "--delete-older-than 150d"; nix.gc.options = "--delete-older-than 150d";

View File

@@ -15,26 +15,27 @@ in
"deepseek-r1:1.5b" "deepseek-r1:1.5b"
"deepseek-r1:32b" "deepseek-r1:32b"
"deepseek-r1:70b" "deepseek-r1:70b"
#"qwen3" "qwen3"
#"qwen3.5:latest" #"qwen3-coder-next"
"qwen3-coder-next" "qwen3-coder-next"
"lennyerik/zeta" "lennyerik/zeta"
"qwen2.5-coder:14b"
"qwen2.5-coder:32b"
"nomic-embed-text:latest" "nomic-embed-text:latest"
"llama4:scout"
"mistral:7b"
"minimax-m2.7:cloud"
"lfm2:24b" "lfm2:24b"
"glm-4.7-flash" "glm-4.7-flash"
"nemotron-cascade-2:30b"
"magistral"
"devstral-small-2"
"starcoder2:15b"
]; ];
models = vars.primary_ollama; models = vars.primary_ollama;
environmentVariables = { environmentVariables = {
FLASH_ATTENTION = "1"; FLASH_ATTENTION = "1";
OLLAMA_KV_CACHE_TYPE = "q4_0"; OLLAMA_KV_CACHE_TYPE = "q8_0";
# Ollama memory configuration # Ollama memory configuration
OLLAMA_MAX_LOADED_MODELS = "3"; OLLAMA_MAX_LOADED_MODELS = "3";
OLLAMA_MAX_QUEUE = "512"; OLLAMA_MAX_QUEUE = "512";
OLLAMA_NUM_PARALLEL = "1"; OLLAMA_NUM_PARALLEL = "3";
# ROCm memory optimization # ROCm memory optimization
#HIP_VISIBLE_DEVICES = "0"; #HIP_VISIBLE_DEVICES = "0";

View File

@@ -31,6 +31,7 @@
grim grim
htop htop
hwloc hwloc
ipmiview
iperf3 iperf3
# ipscan # ipscan
jp2a jp2a

View File

@@ -63,81 +63,8 @@
"latex" "latex"
"terraform" "terraform"
"log" "log"
"context7-mcp-server"
"github-mcp-server"
]; ];
userSettings = { userSettings = {
context_servers = {
nixos = {
command = "nix";
args = [
"run"
"github:utensils/mcp-nixos"
"--"
];
};
};
language_models = {
ollama = {
api_url = "http://192.168.76.2:11434";
context_window = 128000;
# global keep alive doesnt work
#keep_alive = "15m";
available_models = [
{
name = "deepseek-r1:1.5b";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "deepseek-r1:32b";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "deepseek-r1:70b";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "qwen3-coder-next";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "lennyerik/zeta";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "nomic-embed-text:latest";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "lfm2:24b";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "glm-4.7-flash";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "nemotron-cascade-2:30b";
max_tokens = 128000;
keep_alive = "15m";
}
{
name = "magistral";
max_tokens = 128000;
keep_alive = "15m";
}
];
};
};
colorize_brackets = true; colorize_brackets = true;
hard_tabs = false; hard_tabs = false;
vim_mode = true; vim_mode = true;
@@ -150,7 +77,7 @@
agent = { agent = {
default_model = { default_model = {
provider = "ollama"; provider = "ollama";
model = "glm-4.7-flash"; model = "qwen2.5-coder:latest";
}; };
favorite_models = [ ]; favorite_models = [ ];
model_parameters = [ ]; model_parameters = [ ];
@@ -162,16 +89,13 @@
journal = { journal = {
hour_format = "hour24"; hour_format = "hour24";
}; };
edit_predictions = { edit_preditions = {
provider = "ollama"; provider = "open_ai_compatible_api";
ollama = { open_ai_compatible_api = {
#api_url = "http://192.168.76.2:11434/v1/completions"; api_url = "http://localhost:11434/v1/completions";
api_url = "http://192.168.76.2:11434"; model = "zeta:latest";
context_window = 128000; prompt_format = "infer";
model = "lennyerik/zeta";
prompt_format = "qwen";
max_requests = 64; max_requests = 64;
max_output_tokens = 256;
}; };
}; };
texlab = { texlab = {
@@ -220,8 +144,6 @@
# markdown # markdown
nodePackages.markdownlint-cli nodePackages.markdownlint-cli
# insert essential rust dependencies
# doom emacs dependencies # doom emacs dependencies
yaml-language-server yaml-language-server
nodePackages.typescript-language-server nodePackages.typescript-language-server
@@ -262,6 +184,5 @@
# arch zed deps # arch zed deps
nixd nixd
uv
]; ];
} }