Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.openmemind.com/llms.txt

Use this file to discover all available pages before exploring further.

Chat GPT Image 2026年5月8日 21 01 56
The Memind architecture is centered around Memind Core Runtime. Applications and agents can use Memind directly through the Java SDK, or access the same runtime through memind-server. Memind UI sits beside the runtime as an inspection and operation console. At a high level, the system has six major parts:
PartRole
Applications and entry pointsAI agents, Java applications, agent integrations, memind-server, and Memind UI connect to Memind.
Memind Core RuntimeThe central runtime that owns memory writing, context commit, retrieval, runtime configuration, and lifecycle.
Memory Construction EngineThe write path that turns raw input into layered memory.
Cognitive Memory ModelThe memory layers built by Memind: Raw Data, Memory Items, Item Graph, Memory Threads, and Insight Tree.
Memory Retrieval EngineThe read path that searches across memory layers and returns concise context.
Storage, indexing, and model servicesPersistent storage, buffers, vector indexes, text search, LLMs, embeddings, rerankers, parsers, and plugins.

Applications and entry points

Memind can be used in multiple ways. Java applications can embed Memind directly as a library through the Java SDK. Agent integrations can connect coding agents or assistant tools to a Memind runtime. memind-server exposes Memind as a standalone service. Memind UI provides an interface for inspecting and operating the memory system. All of these entry points eventually connect to the same core idea: a Memory runtime that can write memory, commit context, retrieve memory, and build agent context.

Memind Core Runtime

Memind Core Runtime is the center of the system. It contains four main areas:
Runtime areaPurpose
Memory APIProvides the main operations for writing memory, committing buffered context, retrieving memory, and building context windows.
Runtime ConfigurationControls extraction behavior, retrieval behavior, memory-thread behavior, and runtime settings.
Memory Construction EngineHandles the write path from raw input to structured and consolidated memory.
Memory Retrieval EngineHandles the read path from user or agent query to retrieved context.
The runtime is the same whether Memind is embedded in a Java application or hosted by memind-server.

Memory Construction Engine

The Memory Construction Engine is the write path. It progressively transforms raw fragmented information into structured, connected, and consolidated memory.
Raw Input
  -> Context Buffer
  -> Raw Data Layer
  -> Memory Item Layer
  -> Item Graph
  -> Memory Threads
  -> Insight Tree
Raw input can come from conversations, files, URLs, tool output, or custom application content. The context buffer helps chat-style applications accumulate pending and recent messages, then commit conversation segments when a boundary is detected. The Raw Data layer preserves source context through source segments, captions, metadata, and references. The Memory Item layer extracts durable facts, preferences, events, directives, tool experience, playbooks, and resolutions. The Item Graph connects those items through entities, aliases, mentions, semantic links, temporal links, and co-occurrence. Memory Threads organize related items into evolving topics such as projects, workflows, incidents, and decisions. The Insight Tree then consolidates lower-level memory into LEAF, BRANCH, and ROOT insights. This pipeline is the core reason Memind is more than a flat memory store.

Cognitive Memory Model

The Cognitive Memory Model is the layered memory structure produced by Memind.
Raw Data
  -> Memory Items
  -> Item Graph
  -> Memory Threads
  -> Insight Tree
Each layer adds a different kind of understanding. Raw Data preserves what happened. Memory Items extract what matters. The Item Graph connects fragmented knowledge. Memory Threads follow topics as they evolve. The Insight Tree consolidates patterns into higher- level understanding. This model lets Memind move from scattered information to memory that can mature over time.

Memory Retrieval Engine

The Memory Retrieval Engine is the read path. When an agent asks a question, Memind does not only search a vector index. It can retrieve across multiple memory layers, including insights, memory items, raw-data evidence, graph connections, and memory threads.
User / Agent Query
  -> Retrieval Orchestration
  -> SIMPLE Retrieval or DEEP Retrieval
  -> Result Fusion
  -> Context Result
Retrieval orchestration handles admission, cache, query condensation, and strategy selection. SIMPLE retrieval is optimized for low-latency recall. It can combine vector search, keyword search, temporal signals, graph assist, thread assist, and result fusion. DEEP retrieval is designed for harder queries. It can use LLM-assisted retrieval, query expansion, sufficiency checking, graph exploration, thread exploration, and optional reranking. The final context result can include relevant insights, memory items, raw-data evidence, source references, and retrieval trace information.

Storage and indexing layer

Memind separates runtime memory from storage and indexes. The storage and indexing layer supports both construction and retrieval:
ComponentRole
Memory StorePersists raw data, memory items, insights, graph data, thread projections, and resources.
Memory BufferHolds pending conversations, recent conversations, and insight build buffers.
Vector IndexStores embeddings for raw data, memory items, and insights.
Text Search IndexSupports keyword and BM25-style search.
Insight Build StateTracks state used to construct and reorganize the Insight Tree.
Memory Construction writes into this layer. Memory Retrieval reads from it.

Model and extension services

Memind uses model services and extension points around the core runtime. The LLM or structured chat client is used for tasks such as caption generation, memory item extraction, insight generation, commit detection, query expansion, sufficiency checking, and thread enrichment. The embedding model vectorizes raw data, memory items, and insights. An optional reranker can improve deep retrieval ranking. Raw Data plugins, content parsers, and resource fetchers extend what Memind can ingest and how content enters the memory construction pipeline. These services support the runtime, but they are not the memory system by themselves. The memory system is the layered runtime that organizes their outputs into durable, connected, and inspectable memory.

Inspection and operations

Memind UI makes the memory runtime inspectable. It lets developers inspect dashboard state, buffers, Raw Data, Memory Items, the Item Graph, Memory Threads, the Insight Tree, settings, and retrieval traces. This is important because agent memory should not be a black box. Developers need to see what was stored, how it was extracted, how it was connected, what was consolidated, and why certain context was retrieved.

Why this architecture matters

Memind is designed around one principle:
Memory is constructed, connected, consolidated, and evolved.
A simple vector store can retrieve similar text. Memind is built to preserve source context, extract structured memory, connect related knowledge, follow long-running topics, build higher-level insights, and return useful context to agents. That is why the center of the architecture is not a database or an index. It is the Memind Core Runtime.