Renames: - platform/ → workspace-server/ (Go module path stays as "platform" for external dep compat — will update after plugin module republish) - workspace-template/ → workspace/ Removed (moved to separate repos or deleted): - PLAN.md — internal roadmap (move to private project board) - HANDOFF.md, AGENTS.md — one-time internal session docs - .claude/ — gitignored entirely (local agent config) - infra/cloudflare-worker/ → Molecule-AI/molecule-tenant-proxy - org-templates/molecule-dev/ → standalone template repo - .mcp-eval/ → molecule-mcp-server repo - test-results/ — ephemeral, gitignored Security scrubbing: - Cloudflare account/zone/KV IDs → placeholders - Real EC2 IPs → <EC2_IP> in all docs - CF token prefix, Neon project ID, Fly app names → redacted - Langfuse dev credentials → parameterized - Personal runner username/machine name → generic Community files: - CONTRIBUTING.md — build, test, branch conventions - CODE_OF_CONDUCT.md — Contributor Covenant 2.1 All Dockerfiles, CI workflows, docker-compose, railway.toml, render.yaml, README, CLAUDE.md updated for new directory names. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
132 lines
5.4 KiB
Python
132 lines
5.4 KiB
Python
"""Memory consolidation loop.
|
|
|
|
When an agent is idle (no active tasks for a configurable period),
|
|
the consolidation loop wakes up and summarizes noisy local memory
|
|
entries into dense, high-value knowledge facts.
|
|
|
|
Similar to human sleep consolidation — raw scratchpad entries get
|
|
compressed into reusable knowledge.
|
|
"""
|
|
|
|
import asyncio
|
|
import logging
|
|
import os
|
|
|
|
import httpx
|
|
|
|
from platform_auth import auth_headers
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
PLATFORM_URL = os.environ.get("PLATFORM_URL", "http://platform:8080")
|
|
WORKSPACE_ID = os.environ.get("WORKSPACE_ID", "")
|
|
CONSOLIDATION_INTERVAL = float(os.environ.get("CONSOLIDATION_INTERVAL", "300")) # 5 min
|
|
CONSOLIDATION_THRESHOLD = int(os.environ.get("CONSOLIDATION_THRESHOLD", "10")) # min memories before consolidating
|
|
|
|
|
|
class ConsolidationLoop:
|
|
"""Background loop that consolidates local memories when idle."""
|
|
|
|
def __init__(self, agent=None):
|
|
self.agent = agent
|
|
self._running = False
|
|
|
|
async def start(self):
|
|
"""Start the consolidation loop."""
|
|
self._running = True
|
|
logger.info("Memory consolidation loop started (interval=%ss, threshold=%d)",
|
|
CONSOLIDATION_INTERVAL, CONSOLIDATION_THRESHOLD)
|
|
|
|
while self._running:
|
|
await asyncio.sleep(CONSOLIDATION_INTERVAL)
|
|
|
|
if not self._running:
|
|
break
|
|
|
|
try:
|
|
await self._consolidate()
|
|
except Exception as e:
|
|
logger.warning("Consolidation error: %s", e)
|
|
|
|
async def _consolidate(self):
|
|
"""Check if consolidation is needed and run it."""
|
|
async with httpx.AsyncClient(timeout=10.0) as client:
|
|
# Fetch local memories
|
|
resp = await client.get(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories",
|
|
params={"scope": "LOCAL"},
|
|
headers=auth_headers(),
|
|
)
|
|
if resp.status_code != 200:
|
|
return
|
|
|
|
memories = resp.json()
|
|
if len(memories) < CONSOLIDATION_THRESHOLD:
|
|
return
|
|
|
|
logger.info("Consolidating %d local memories", len(memories))
|
|
|
|
# Build a summary of all local memories
|
|
contents = [m["content"] for m in memories]
|
|
summary_prompt = (
|
|
"Summarize the following workspace memories into 3-5 key facts. "
|
|
"Each fact should be a single, clear sentence capturing the most "
|
|
"important and reusable knowledge:\n\n"
|
|
+ "\n".join(f"- {c}" for c in contents)
|
|
)
|
|
|
|
# Use the agent to generate the summary if available
|
|
summary = ""
|
|
if self.agent:
|
|
try:
|
|
result = await self.agent.ainvoke(
|
|
{"messages": [("user", summary_prompt)]},
|
|
config={"configurable": {"thread_id": "consolidation"}},
|
|
)
|
|
messages = result.get("messages", [])
|
|
summary = ""
|
|
for msg in reversed(messages):
|
|
content = getattr(msg, "content", "")
|
|
if isinstance(content, str) and content.strip():
|
|
msg_type = getattr(msg, "type", "")
|
|
if msg_type != "human":
|
|
summary = content
|
|
break
|
|
|
|
if summary:
|
|
# Store consolidated summary as a TEAM memory — only delete originals if POST succeeds
|
|
resp = await client.post(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories",
|
|
json={"content": f"[Consolidated] {summary}", "scope": "TEAM"},
|
|
headers=auth_headers(),
|
|
)
|
|
if resp.status_code in (200, 201):
|
|
# Safe to delete originals — consolidated version is saved
|
|
for m in memories:
|
|
await client.delete(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories/{m['id']}",
|
|
headers=auth_headers(),
|
|
)
|
|
logger.info("Consolidated %d memories into team knowledge", len(memories))
|
|
else:
|
|
logger.warning("Consolidation POST failed (status %d) — keeping originals", resp.status_code)
|
|
except Exception as e:
|
|
logger.error(
|
|
"CONSOLIDATION: Agent summarization failed (rate limit? model error?): %s. "
|
|
"Falling back to simple concatenation.", e
|
|
)
|
|
# Fall through to concatenation below
|
|
|
|
# Fallback: concatenate without agent summarization
|
|
if not (self.agent and summary):
|
|
combined = " | ".join(contents[:20])
|
|
await client.post(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories",
|
|
json={"content": f"[Consolidated] {combined}", "scope": "TEAM"},
|
|
headers=auth_headers(),
|
|
)
|
|
logger.info("Consolidated %d memories via concatenation fallback", len(memories))
|
|
|
|
def stop(self):
|
|
self._running = False
|