Back to Skills

localai

verified

LocalAI local inference API management via Podman Quadlet. Provides an OpenAI-compatible API for local model inference with GPU acceleration. Use when users need to configure, start, or manage the LocalAI service.

View on GitHub

Marketplace

bazzite-ai-plugins

atrawog/bazzite-ai-plugins

Plugin

bazzite-ai

productivity

Repository

atrawog/bazzite-ai-plugins

bazzite-ai/skills/localai/SKILL.md

Last Verified

January 21, 2026

Install Skill

Select agents to install to:

Scope:
npx add-skill https://github.com/atrawog/bazzite-ai-plugins/blob/main/bazzite-ai/skills/localai/SKILL.md -a claude-code --skill localai

Installation paths:

Claude
.claude/skills/localai/
Powered by add-skill CLI

Instructions

# LocalAI - Local AI Inference API

## Overview

The `localai` command manages the LocalAI service using Podman Quadlet containers. It provides an OpenAI-compatible API for running AI models locally with GPU acceleration.

**Key Features:**

- OpenAI-compatible API endpoints
- GPU-specific container images (auto-selected)
- Multiple GPU support (NVIDIA, AMD, Intel)
- Cross-pod DNS via `bazzite-ai` network

## Quick Reference

| Action | Command | Description |
|--------|---------|-------------|
| Config | `ujust localai config` | Configure LocalAI |
| Delete | `ujust localai delete` | Remove instance config and container |
| Logs | `ujust localai logs [--lines=N]` | View container logs |
| Restart | `ujust localai restart` | Restart server |
| Shell | `ujust localai shell [-- CMD]` | Open shell or execute command in container |
| Start | `ujust localai start` | Start LocalAI server |
| Status | `ujust localai status` | Show instance status |
| Stop | `ujust localai stop` | Stop LocalAI server |
| URL | `ujust localai url` | Show OpenAI-compatible API URL |

## Parameters

| Parameter | Long Flag | Short | Default | Description |
|-----------|-----------|-------|---------|-------------|
| Port | `--port` | `-p` | `8080` | Host port for API |
| Image | `--image` | `-i` | (auto by GPU) | Container image |
| Tag | `--tag` | `-t` | `latest` | Image tag |
| Bind | `--bind` | `-b` | `127.0.0.1` | Bind address |
| Config Dir | `--config-dir` | `-c` | `~/.config/localai/1` | Config/models directory |
| Workspace | `--workspace-dir` | `-w` | (empty) | Workspace mount |
| GPU Type | `--gpu-type` | `-g` | `auto` | GPU type |
| Instance | `--instance` | `-n` | `1` | Instance number or `all` |
| Lines | `--lines` | `-l` | `50` | Log lines to show |

## GPU-Specific Images

LocalAI uses different container images optimized for each GPU type:

| GPU Type | Image | Auto-Selected? |
|----------|-------|----------------|
| CPU (none) | `localai/localai:latest` | Yes |
| NVIDIA | `loca

Validation Details

Front Matter
Required Fields
Valid Name Format
Valid Description
Has Sections
Allowed Tools
Instruction Length:
8239 chars