"""Local AI provider — bundled llama-server (OpenAI-compatible API).""" from __future__ import annotations # TODO: Implement local provider # - Connect to llama-server on localhost:{port} # - Use openai SDK with custom base_url # - Support chat and streaming