rebuilding nix files

This commit is contained in:
2026-02-28 11:31:52 +01:00
parent f347052173
commit 4135e7df42
3 changed files with 339 additions and 324 deletions
+317 -312
View File
File diff suppressed because it is too large Load Diff
+11 -6
View File
@@ -1680,13 +1680,15 @@ in
MISTRAL_API_KEY = envVars.MISTRAL_API_KEY or "";
};
# Configure Ollama service with explicit type declarations
# Configure Ollama service using the correct Home Manager syntax
systemd.user.services.ollama = {
description = "Ollama service for local AI models";
wantedBy = [ "default.target" ];
after = lib.mkForce [ "network.target" ]; # Explicitly force as a list
Unit = {
Description = "Ollama service for local AI models";
After = [ "network.target" ];
Wants = [ "network.target" ];
};
serviceConfig = {
Service = {
Type = "forking";
ExecStart = "${pkgs.ollama-vulkan}/bin/ollama serve";
ExecStartPost = ''
@@ -1696,6 +1698,10 @@ in
'';
Restart = "on-failure";
};
Install = {
WantedBy = [ "default.target" ];
};
};
# ZED configuration
@@ -1712,7 +1718,6 @@ in
}
);
# --- Usage Notes ---
# 1. Pulling Additional Models:
# To add more models later, run:
+11 -6
View File
@@ -35,13 +35,15 @@ in
MISTRAL_API_KEY = envVars.MISTRAL_API_KEY or "";
};
# Configure Ollama service with explicit type declarations
# Configure Ollama service using the correct Home Manager syntax
systemd.user.services.ollama = {
description = "Ollama service for local AI models";
wantedBy = [ "default.target" ];
after = lib.mkForce [ "network.target" ]; # Explicitly force as a list
Unit = {
Description = "Ollama service for local AI models";
After = [ "network.target" ];
Wants = [ "network.target" ];
};
serviceConfig = {
Service = {
Type = "forking";
ExecStart = "${pkgs.ollama-vulkan}/bin/ollama serve";
ExecStartPost = ''
@@ -51,6 +53,10 @@ in
'';
Restart = "on-failure";
};
Install = {
WantedBy = [ "default.target" ];
};
};
# ZED configuration
@@ -67,7 +73,6 @@ in
}
);
# --- Usage Notes ---
# 1. Pulling Additional Models:
# To add more models later, run: