📢 Announcing our research paper: Zentry achieves 26% higher accuracy than OpenAI Memory, 91% lower latency, and 90% token savings! Read the paper to learn how we're revolutionizing AI agent memory.
Usage
Running Completely Locally
You can also use LM Studio for both LLM and embedding to run Zentry entirely locally:When using LM Studio for both LLM and embedding, make sure you have:
- An LLM model loaded for generating responses
- An embedding model loaded for vector embeddings
- The server enabled with the correct endpoints accessible
To use LM Studio, you need to:
- Download and install LM Studio
- Start a local server from the “Server” tab
- Set the appropriate
lmstudio_base_urlin your configuration (default is usually http://localhost:1234/v1)
Config
All available parameters for thelmstudio config are present in Master List of All Params in Config.