- Replace Ollama dependency with bundled llama-server (llama.cpp)
so users need no separate install for local AI inference
- Rust backend manages llama-server lifecycle (spawn, port, shutdown)
- Add MIT license for open source release
- Update architecture doc, CLAUDE.md, and README accordingly
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>