Back to Skills

axiom-ios-ml

verified

Use when deploying ANY machine learning model on-device, converting models to CoreML, compressing models, or implementing speech-to-text. Covers CoreML conversion, MLTensor, model compression (quantization/palettization/pruning), stateful models, KV-cache, multi-function models, async prediction, SpeechAnalyzer, SpeechTranscriber.

View on GitHub

Marketplace

axiom-marketplace

CharlesWiltgen/Axiom

Plugin

axiom

Repository

CharlesWiltgen/Axiom
289stars

.claude-plugin/plugins/axiom/skills/axiom-ios-ml/SKILL.md

Last Verified

January 16, 2026

Install Skill

Select agents to install to:

Scope:
npx add-skill https://github.com/CharlesWiltgen/Axiom/blob/main/.claude-plugin/plugins/axiom/skills/axiom-ios-ml/SKILL.md -a claude-code --skill axiom-ios-ml

Installation paths:

Claude
.claude/skills/axiom-ios-ml/
Powered by add-skill CLI

Instructions

# iOS Machine Learning Router

**You MUST use this skill for ANY on-device machine learning or speech-to-text work.**

## When to Use

Use this router when:
- Converting PyTorch/TensorFlow models to CoreML
- Deploying ML models on-device
- Compressing models (quantization, palettization, pruning)
- Working with large language models (LLMs)
- Implementing KV-cache for transformers
- Using MLTensor for model stitching
- Building speech-to-text features
- Transcribing audio (live or recorded)

## Routing Logic

### CoreML Work

**Implementation patterns** → `/skill coreml`
- Model conversion workflow
- MLTensor for model stitching
- Stateful models with KV-cache
- Multi-function models (adapters/LoRA)
- Async prediction patterns
- Compute unit selection

**API reference** → `/skill coreml-ref`
- CoreML Tools Python API
- MLModel lifecycle
- MLTensor operations
- MLComputeDevice availability
- State management APIs
- Performance reports

**Diagnostics** → `/skill coreml-diag`
- Model won't load
- Slow inference
- Memory issues
- Compression accuracy loss
- Compute unit problems

### Speech Work

**Implementation patterns** → `/skill speech`
- SpeechAnalyzer setup (iOS 26+)
- SpeechTranscriber configuration
- Live transcription
- File transcription
- Volatile vs finalized results
- Model asset management

## Decision Tree

```
User asks about on-device ML or speech
  ├─ Machine learning?
  │   ├─ Implementing/converting? → coreml
  │   ├─ Need API reference? → coreml-ref
  │   └─ Debugging issues? → coreml-diag
  └─ Speech-to-text?
      └─ Any speech work → speech
```

## Critical Patterns

**coreml**:
- Model conversion (PyTorch → CoreML)
- Compression (palettization, quantization, pruning)
- Stateful KV-cache for LLMs
- Multi-function models for adapters
- MLTensor for pipeline stitching
- Async concurrent prediction

**coreml-diag**:
- Load failures and caching
- Inference performance issues
- Memory pressure from models
- Accuracy degradation from compression

**spe

Validation Details

Front Matter
Required Fields
Valid Name Format
Valid Description
Has Sections
Allowed Tools
Instruction Length:
2728 chars