Getting Started
Get momo-kibidango running in your OpenClaw environment in under 5 minutes.
Prerequisites
- ✓ macOS with Apple Silicon (M1/M2/M3/M4)
- ✓ OpenClaw v1.12.0 or higher
- ✓ 16GB RAM minimum (32GB recommended)
- ✓ 50GB free disk space
Quick Install
Terminal
pip install momo-kibidangoOpenClaw Integration
Add momo-kibidango to your OpenClaw configuration:
~/.openclaw/config.json
{
"models": {
"accelerated": {
"provider": "momo-kibidango",
"base_model": "sonnet-3.5",
"draft_models": ["haiku-2", "haiku-3"],
"device": "mps"
}
}
}Verify Installation
Test that momo-kibidango is working correctly:
Terminal
momo-kibidango testExpected output:
✅ Models loaded successfully
✅ Draft model: haiku-2 (45.6 tok/s)
✅ Middle model: haiku-3 (30.5 tok/s)
✅ Target model: sonnet-3.5 (12.5 tok/s)
✅ Acceleration: 1.97x (24.6 tok/s effective)
✅ Memory usage: 11.6GBFirst Run
Use momo-kibidango with your OpenClaw subagents:
Python
from openclaw import spawn_subagent
# Automatically uses accelerated inference
result = spawn_subagent(
task="Write a comprehensive analysis",
model="accelerated" # Uses momo-kibidango
)💡 Pro Tip
The first run will download and cache models (~15GB). This is a one-time operation. Subsequent runs will start in under 10 seconds.
Monitor Performance
View real-time metrics with the built-in dashboard:
Terminal
momo-kibidango monitorThis opens a web dashboard at http://localhost:9090 showing:
- • Token generation speed
- • Acceptance rates per model
- • Memory usage
- • Cache hit rates
- • Fallback statistics
🎉 You're Ready!
momo-kibidango is now accelerating your LLM inference. Check out the configuration guide to fine-tune performance for your specific use case.