Version0.1.4
Revision6
Size226.6 MB
Licenseunset
Confinementstrict
Basecore24

OpenMemory Snap.

ScreenshotScreenshotScreenshotScreenshot

A snap for OpenMemory,

OpenMemory is a local memory infrastructure powered by Mem0 that lets you carry your memory across any AI app. It provides a unified memory layer that stays with you, enabling agents and assistants to remember what matters across applications.

docs: https://docs.mem0.ai/openmemory/overview

OpenMemory can be utilized entirely locally by leveraging Ollama for both the embedding model and the language model (LLM).

Download: Ollama https://ollama.com/download

llm

  "provider": "ollama",
  "config": {
       "model": "llama3.1:latest",
       "temperature": 0,
       "maxtokens": 2000,
       "ollamabaseurl": "http://localhost:11434",  # Ensure this URL is correct
  }

embedder

  "provider": "ollama",
  "config": {
       "model": "nomic-embed-text:latest",
       "embeddingdims": 768,
       "ollamabaseurl": "http://localhost:11434",
  }


The OpenMemory with Ollama is a private, local-first memory server that creates a shared, persistent memory layer for your MCP-compatible tools.
This runs entirely on your machine, enabling seamless context handoff across tools. Whether you're switching between development, planning, or debugging environments, your AI assistants can access relevant memory without needing repeated instructions.

Powered by

https://mem0.ai

https://ollama.com

https://qdrant.tech/

https://tauri.app/

Update History

0.1.4 (6)
13 Dec 2025, 09:47 UTC

Published13 Nov 2025, 11:31 UTC

Last updated20 Nov 2025, 17:08 UTC

First seen13 Dec 2025, 09:47 UTC