clawhub/ollama-memory-embeddings
Configure OpenClaw memory search to use Ollama as the embeddings server (OpenAI-compatible /v1/embeddings) instead of the built-in node-llama-cpp local GGUF loading. Includes interactive model selection and optional import of an existing local embedding GGUF into Ollama.
Risk Score
0
out of 100
Findings by Severity (Latest Scan)
CodeThreat AppSec
Full SAST + SCA agentic security analysis for MCP servers and Skills.