No description
Find a file
Kent Overstreet 15d4bfa01f observation: chunk large transcripts, remove format_segment limit
Large conversation segments are now split into 50KB chunks with 10KB
overlap, instead of being truncated to 8000 chars (which was broken
anyway — broke after exceeding, not before). Each chunk gets its own
candidate ID for independent mining and dedup.

format_segment simplified: no size limit, added timestamps to output
so observation agent can cross-reference with journal entries.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-03-16 20:52:20 -04:00
.cargo daemon: resource-gated scheduling, fact-mine integration, systemd 2026-03-05 15:31:08 -05:00
.claude stash DMN algorithm plan and connector prompt fix 2026-03-05 10:24:24 -05:00
doc move README.md to toplevel 2026-03-05 15:55:59 -05:00
docs experience-mine: per-segment dedup keys, retry backoff 2026-03-09 02:27:51 -04:00
jobkit-daemon extract jobkit-daemon library from poc-memory daemon 2026-03-14 02:40:30 -04:00
poc-daemon poc-daemon: fix idle nudge and notification delivery 2026-03-16 17:09:27 -04:00
poc-memory observation: chunk large transcripts, remove format_segment limit 2026-03-16 20:52:20 -04:00
prompts experience-mine: link at creation time, remove # from new keys 2026-03-14 16:25:31 -04:00
.gitignore knowledge agents: extractor, connector, challenger, observation 2026-03-03 10:56:44 -05:00
Cargo.lock evaluate: switch to Elo ratings with skillratings crate 2026-03-14 19:53:46 -04:00
Cargo.toml extract jobkit-daemon library from poc-memory daemon 2026-03-14 02:40:30 -04:00
README.md docs: expand README getting started section 2026-03-07 13:58:19 -05:00

poc-memory

A persistent memory and notification system for AI assistants, modelled after the human hippocampus. Combines episodic memory (timestamped journal of experiences) with an associative knowledge graph (weighted nodes connected by typed relations), and layered background processes that maintain graph health — mirroring how biological memory consolidates during rest.

Components

Component What it does Docs
Memory store Knowledge graph with episodic journal, TF-IDF search, spectral embedding, weight decay docs/memory.md
Memory daemon Background pipeline: experience-mine, fact-mine, consolidation docs/daemon.md
Notification daemon Activity-aware message routing from IRC and Telegram docs/notifications.md
Hooks Claude Code integration: memory recall and notification delivery docs/hooks.md

Getting started

Install

cargo install --path .

This builds four binaries:

  • poc-memory — memory store CLI (search, journal, consolidation)
  • memory-search — Claude Code hook for memory recall
  • poc-daemon — notification daemon (IRC, Telegram, idle tracking)
  • poc-hook — Claude Code hook for session lifecycle events

Initialize

poc-memory init

Creates the store at ~/.claude/memory/nodes.capnp and a default config at ~/.config/poc-memory/config.jsonl. Edit the config to set your name, configure context groups, and point at your projects directory.

Set up hooks

Add to ~/.claude/settings.json (see docs/hooks.md for full details):

{
  "hooks": {
    "UserPromptSubmit": [{"hooks": [
      {"type": "command", "command": "memory-search", "timeout": 10},
      {"type": "command", "command": "poc-hook", "timeout": 5}
    ]}],
    "Stop": [{"hooks": [
      {"type": "command", "command": "poc-hook", "timeout": 5}
    ]}]
  }
}

This gives your AI assistant persistent memory across sessions — relevant memories are recalled on each prompt, and experiences are extracted from transcripts after sessions end.

Start the background daemon

poc-memory daemon

The daemon watches for completed session transcripts and automatically extracts experiences and facts into the knowledge graph. See docs/daemon.md for pipeline details and diagnostics.

Basic usage

poc-memory journal-write "learned that X does Y"  # Write to journal
poc-memory search "some topic"                     # Search the graph
poc-memory status                                  # Store overview

For AI assistants

  • Search before creating: poc-memory search before writing new nodes
  • Close the feedback loop: poc-memory used KEY / poc-memory wrong KEY
  • Journal is the river, topic nodes are the delta: write experiences to the journal, pull themes into topic nodes during consolidation
  • Notifications flow automatically: IRC/Telegram messages arrive as additionalContext
  • Use daemon commands directly: poc-daemon irc send #channel msg, poc-daemon telegram send msg