It would be really helpful if Memori supported (and documented) how to connect to local OpenAI-compatible endpoints. ## Motivation Many developers run models locally using projects like [llama.cpp](https://github.com/ggerganov/llama.cpp), [koboldcpp](https://github.com/LostRuins/koboldcpp), or [jan.ai](https://jan.ai/). Some users don’t want to send data to external APIs for privacy or cost reasons. Having a straightforward way to configure Memori with local endpoints would make it more flexible and self-hosting–friendly. ## Proposal Add instructions in the README on how to: - Point Memori to a local LLM endpoint (OpenAI API-compatible). - Configure embeddings with a different base URL than the LLM (since some setups split LLMs and embedding services). ## Benefits - Makes Memori usable in fully local / offline setups. - Supports wider community adoption across different backends. - Increases flexibility for developers working with custom infra.