1
0
mirror of https://github.com/XNM1/linux-nixos-hyprland-config-dotfiles.git synced 2025-09-15 09:45:58 +03:00
Files
linux-nixos-hyprland-config…/nixos/llm.nix
xnm 49f23c4fd4 feat(ai): 🤖 update LLM models and configuration
- Changed default model from `ollama:llama3.2:3b` to `ollama:gemma3:4b`
- Updated model token limits and added new models:
  - Added `gemma3:4b` with vision support
  - Removed `phi4:14b` and added `phi4-reasoning:14b`
  - Updated token limits for existing models
- Added new AI-related roles:
  - `commit-message.md`
  - `email-answer.md`
  - `emoji-commit-message.md`
  - `git-branch.md`
  - `improve-prompt.md`
  - `improve-writing.md`
  - `linkedin-answer.md`
  - Enhanced `to-emoji.md` and added `to-emojies.md`
- Added `ai` alias for `aichat` in fish config
- Updated NixOS configuration to load new models
2025-06-01 18:57:33 +03:00

45 lines
1.0 KiB
Nix

{ pkgs, config, ... }:
{
services.ollama = {
enable = true;
loadModels = [ "llama3.2:3b" "phi4-reasoning:14b" "dolphin3:8b" "smallthinker:3b" "gemma3:4b" "gemma3:12b" "gemma3:27b" "deepcoder:14b" "qwen3:14b" "nomic-embed-text" ];
acceleration = "cuda";
};
services.searx = {
enable = true;
settings = {
server = {
port = 7777;
bind_address = "127.0.0.1";
secret_key = "@SEARX_SECRET_KEY@"; # FIXME: Set up this key in the .env file described below, name of variable `SEARX_SECRET_KEY`
};
search = {
formats = [ "html" "json" ];
};
};
environmentFile = "${config.users.users.xnm.home}/.config/.env.searxng"; # FIXME: The location of the `.env` file where you need to set up the key
};
services.open-webui = {
enable = true;
port = 8888;
host = "127.0.0.1";
};
environment.systemPackages = with pkgs; [
oterm
alpaca
aichat
fabric-ai
aider-chat
# tgpt
# smartcat
# nextjs-ollama-llm-ui
# open-webui
];
}