Skills for using Bazzite AI OS features via ujust commands
Apptainer (Singularity) container management for HPC workloads. Build SIF images, run containers with GPU passthrough. Use when users need HPC-compatible containerization or need to pull/run Apptainer images.
bootc VM management via bcvk (bootc virtualization kit). Run bootable containers as VMs for testing. Supports ephemeral (quick test) and persistent modes. Use when users need to test bootable container images as virtual machines.
ComfyUI node-based Stable Diffusion interface. GPU-accelerated image generation with custom node support and CivitAI model downloads. Use 'ujust comfyui' for configuration, lifecycle management, and model/node operations.
Unified system configuration dispatcher for bazzite-ai. Manages services (Docker, Cockpit, SSH), desktop settings (gamemode, Steam), security (passwordless sudo), and development environment (GPU containers). Use when users need to enable/disable system features or check configuration status.
Development tool installation dispatcher for bazzite-ai. Installs Claude Code, pixi, chunkhound, bcvk, linters, flatpaks, and more. Use when users need to install standalone developer tools (not services with lifecycle management).
Jellyfin media server management via Podman Quadlet. Supports multi-instance deployment, hardware transcoding (NVIDIA/AMD/Intel), and FUSE filesystem mounts. Use when users need to set up or manage Jellyfin media servers.
JupyterLab ML/AI development environment management via Podman Quadlet. Supports multi-instance deployment, GPU acceleration (NVIDIA/AMD/Intel), token authentication, and per-instance configuration. Use when users need to configure, start, stop, or manage JupyterLab containers for ML development.
k3d Kubernetes cluster management - lightweight k3s clusters running in Podman containers on the bazzite-ai network. Supports GPU passthrough, multi-instance, and service discovery with other bazzite-ai pods. Use when users need to run Kubernetes workloads or deploy k8s-based applications locally.
LocalAI local inference API management via Podman Quadlet. Provides an OpenAI-compatible API for local model inference with GPU acceleration. Use when users need to configure, start, or manage the LocalAI service.
Ollama LLM inference server management via Podman Quadlet. Single-instance design with GPU acceleration for running local LLMs. Use when users need to configure Ollama, pull models, run inference, or manage the Ollama server.
Open WebUI AI chat interface management via Podman Quadlet. Provides a web UI for interacting with Ollama models. Use when users need to configure, start, or manage the Open WebUI service.
Aggregate management for all AI pod services. Provides status overview and bulk operations across all pod containers (ollama, jupyter, comfyui, openwebui, localai, fiftyone, jellyfin, runners).
Self-hosted GitHub Actions runner management via Podman Quadlet. Supports multi-instance pools with ephemeral storage, automatic token generation, and rolling updates. Use when users need to set up CI/CD runners for their GitHub repositories.
Tailscale Serve management for exposing local services to your tailnet. Auto-detects running bazzite-ai services and creates persistent HTTPS endpoints. Use when users need to expose Jupyter, Ollama, ComfyUI or other services to their Tailscale network.
Runtime verification tests for bazzite-ai installation. Tests GPU detection, CUDA, PyTorch, service health, network connectivity, and pod lifecycles. Use when users need to verify their bazzite-ai installation works correctly.
QCOW2 virtual machine management using libvirt. Creates VMs from pre-built images downloaded from R2 CDN with cloud-init customization. Supports SSH, VNC, and virtiofs home directory sharing. Use when users need to create, manage, or connect to bazzite-ai VMs.
ML/AI development workflows for JupyterLab - LangChain, RAG, fine-tuning, and model optimization
Direct Preference Optimization for learning from preference pairs. Covers DPOTrainer, preference dataset preparation, implicit reward modeling, and beta tuning for stable preference learning without explicit reward models. Includes thinking quality patterns.
Model fine-tuning with PyTorch and HuggingFace Trainer. Covers dataset preparation, tokenization, training loops, TrainingArguments, SFTTrainer for instruction tuning, evaluation, and checkpoint management. Includes Unsloth recommendations.
Group Relative Policy Optimization for reinforcement learning from human feedback. Covers GRPOTrainer, reward function design, policy optimization, and KL divergence constraints for stable RLHF training. Includes thinking-aware reward patterns.
Fast inference with Unsloth and vLLM backend. Covers model loading, fast_generate(), thinking model output parsing, and memory management for efficient inference.
Parameter-efficient fine-tuning with LoRA and Unsloth. Covers LoraConfig, target module selection, QLoRA for 4-bit training, adapter merging, and Unsloth optimizations for 2x faster training.
Advanced QLoRA experiments and comparisons. Covers alpha scaling, LoRA rank selection, target module strategies, continual learning, multi-adapter hot-swapping, and quantization comparison (4-bit vs BF16).
Model quantization for efficient inference and training. Covers precision types (FP32, FP16, BF16, INT8, INT4), BitsAndBytes configuration, memory estimation, and performance tradeoffs.
Reward model training for RLHF pipelines. Covers RewardTrainer, preference dataset preparation, sequence classification heads, and reward scaling for stable reinforcement learning. Includes thinking quality scoring patterns.
Reinforcement Learning with Leave-One-Out estimation for policy optimization. Covers RLOOTrainer, reward function integration, baseline estimation, and variance reduction techniques for stable RL training. Includes thinking-aware patterns.
Supervised Fine-Tuning with SFTTrainer and Unsloth. Covers dataset preparation, chat template formatting, training configuration, and Unsloth optimizations for 2x faster instruction tuning. Includes thinking model patterns.
Transformer architecture fundamentals. Covers self-attention mechanism, multi-head attention, feed-forward networks, layer normalization, and residual connections. Essential concepts for understanding LLMs.
Vision model fine-tuning with FastVisionModel. Covers Pixtral, Ministral VL training, UnslothVisionDataCollator, image+text datasets, and vision-specific LoRA configuration.
Skills for using Bazzite OS features via ujust commands
Third-party application installation for Bazzite. CoolerControl, DisplayLink, JetBrains Toolbox, OpenRazer, tablet drivers, scrcpy, and more. Use when users need to install hardware-specific or specialized applications.
Audio configuration for Bazzite. Virtual audio channels for Game/Voice/Browser/Music, 7.1 surround for headphones, Bluetooth headset profiles, and PipeWire management. Use when users need to configure audio on Bazzite.
Boot configuration for Bazzite OS. BIOS/UEFI access, GRUB menu settings, secure boot key enrollment, and Windows dual-boot setup. Use when users need to configure boot options or access BIOS settings.
Desktop customization for Bazzite. GTK theme restoration, terminal transparency, and MOTD settings. Use when users need to customize their desktop appearance.
Distrobox container management for Bazzite. Create containers from manifests, custom containers, app-specific containers (brew), and DaVinci Resolve installation. Use when users need to work with distrobox containers.
Gaming ecosystem for Bazzite. Steam fixes, Proton troubleshooting, EmuDeck emulation, Decky Loader plugins, Sunshine game streaming, frame generation, and media apps. Use when users need help with gaming on Bazzite.
GPU driver configuration for Bazzite. NVIDIA proprietary drivers, Optimus laptops, NVK (open-source NVIDIA), GPU switching, Broadcom WiFi, and Mesa testing builds. Use when users need to configure graphics drivers.
Network configuration for Bazzite. iwd WiFi backend, Wake-on-LAN, and Tailscale VPN. Use when users need to configure network services. For SSH, see bazzite-ai:config.
Security configuration for Bazzite. LUKS disk encryption with TPM auto-unlock, secure boot key management, and sudo password feedback. Use when users need to configure security features.
Storage management for Bazzite. Automounting drives (BTRFS/EXT4, Framework, SteamOS), BTRFS deduplication, rmlint disk trimming, and snapper snapshots. Use when users need to configure disk and partition management.
System maintenance for Bazzite OS. Updates via topgrade, cleanup of podman/flatpaks, viewing logs and changelogs, diagnostics, and power measurements. Use when users need to update, clean, or diagnose their Bazzite system.
GPU passthrough and virtualization for Bazzite. KVM/VFIO setup, Looking Glass (kvmfr), USB hotplug for VMs, and libvirt configuration. Use when users need GPU passthrough or advanced virtualization features.
Ollama API operations for LLM inference, embeddings, and model management
Development tools and enforcement agents for Bazzite AI contributors
Development: Unified build system for OS images, pods, VMs, and ISOs. Run from repository root with 'just build <subcommand>'. Includes smart cache strategy that matches GitHub Actions for optimal build times.
Development: Cleanup and maintenance for the development environment. Removes build artifacts, caches, containers, and recovers disk space. Run from repository root with 'just clean'. Use when developers need to free disk space or reset the build environment.
Overlay testing session management for bazzite-ai development. Enables live editing of justfiles via symlinks to /usr on immutable OS (OSTree) or traditional Linux systems. Use when users need to test ujust changes, enable overlay mode, troubleshoot testing sessions, or run VM/install tests.