IMPACT WITHOUT THIS FIX: deploying PR #31 (WorkspaceAuth middleware on /workspaces/*) without this patch causes EVERY delegation cycle to silently break — the heartbeat poll returns 401, the self-message A2A POST returns 401, agents never wake up after task completion, and memory consolidation stops. The entire multi-agent coordination system degrades to single-shot interactions with no result delivery. Changes (all using the existing platform_auth.auth_headers() pattern already used for POST /registry/heartbeat): heartbeat.py — 5 calls fixed: - GET /workspaces/:id/delegations (delegation poll) - GET /workspaces/:id (self workspace info for parent lookup) - GET /workspaces/{parent_id} (parent workspace name lookup) - POST /workspaces/:id/a2a (self-message to wake agent on results) - POST /workspaces/:id/notify (canvas delegation result notification) Also moved `from platform_auth import auth_headers` from inline (per-call) to module-level import so _check_delegations() can use it without re-importing. consolidation.py — 4 calls fixed: - GET /workspaces/:id/memories (fetch memories for consolidation) - POST /workspaces/:id/memories (write consolidated summary — agent path) - DELETE /workspaces/:id/memories/:id (delete original memories post-consolidation) - POST /workspaces/:id/memories (write consolidated summary — fallback path) a2a_client.py — 1 call fixed: - GET /workspaces/:id (get_workspace_info()) ⚠️ DEPLOYMENT NOTE: This PR MUST be merged and deployed at the same time as PR #31 (WorkspaceAuth middleware). Deploying #31 without this fix will immediately break all delegation result delivery. Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
132 lines
5.4 KiB
Python
132 lines
5.4 KiB
Python
"""Memory consolidation loop.
|
|
|
|
When an agent is idle (no active tasks for a configurable period),
|
|
the consolidation loop wakes up and summarizes noisy local memory
|
|
entries into dense, high-value knowledge facts.
|
|
|
|
Similar to human sleep consolidation — raw scratchpad entries get
|
|
compressed into reusable knowledge.
|
|
"""
|
|
|
|
import asyncio
|
|
import logging
|
|
import os
|
|
|
|
import httpx
|
|
|
|
from platform_auth import auth_headers
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
PLATFORM_URL = os.environ.get("PLATFORM_URL", "http://platform:8080")
|
|
WORKSPACE_ID = os.environ.get("WORKSPACE_ID", "")
|
|
CONSOLIDATION_INTERVAL = float(os.environ.get("CONSOLIDATION_INTERVAL", "300")) # 5 min
|
|
CONSOLIDATION_THRESHOLD = int(os.environ.get("CONSOLIDATION_THRESHOLD", "10")) # min memories before consolidating
|
|
|
|
|
|
class ConsolidationLoop:
|
|
"""Background loop that consolidates local memories when idle."""
|
|
|
|
def __init__(self, agent=None):
|
|
self.agent = agent
|
|
self._running = False
|
|
|
|
async def start(self):
|
|
"""Start the consolidation loop."""
|
|
self._running = True
|
|
logger.info("Memory consolidation loop started (interval=%ss, threshold=%d)",
|
|
CONSOLIDATION_INTERVAL, CONSOLIDATION_THRESHOLD)
|
|
|
|
while self._running:
|
|
await asyncio.sleep(CONSOLIDATION_INTERVAL)
|
|
|
|
if not self._running:
|
|
break
|
|
|
|
try:
|
|
await self._consolidate()
|
|
except Exception as e:
|
|
logger.warning("Consolidation error: %s", e)
|
|
|
|
async def _consolidate(self):
|
|
"""Check if consolidation is needed and run it."""
|
|
async with httpx.AsyncClient(timeout=10.0) as client:
|
|
# Fetch local memories
|
|
resp = await client.get(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories",
|
|
params={"scope": "LOCAL"},
|
|
headers=auth_headers(),
|
|
)
|
|
if resp.status_code != 200:
|
|
return
|
|
|
|
memories = resp.json()
|
|
if len(memories) < CONSOLIDATION_THRESHOLD:
|
|
return
|
|
|
|
logger.info("Consolidating %d local memories", len(memories))
|
|
|
|
# Build a summary of all local memories
|
|
contents = [m["content"] for m in memories]
|
|
summary_prompt = (
|
|
"Summarize the following workspace memories into 3-5 key facts. "
|
|
"Each fact should be a single, clear sentence capturing the most "
|
|
"important and reusable knowledge:\n\n"
|
|
+ "\n".join(f"- {c}" for c in contents)
|
|
)
|
|
|
|
# Use the agent to generate the summary if available
|
|
summary = ""
|
|
if self.agent:
|
|
try:
|
|
result = await self.agent.ainvoke(
|
|
{"messages": [("user", summary_prompt)]},
|
|
config={"configurable": {"thread_id": "consolidation"}},
|
|
)
|
|
messages = result.get("messages", [])
|
|
summary = ""
|
|
for msg in reversed(messages):
|
|
content = getattr(msg, "content", "")
|
|
if isinstance(content, str) and content.strip():
|
|
msg_type = getattr(msg, "type", "")
|
|
if msg_type != "human":
|
|
summary = content
|
|
break
|
|
|
|
if summary:
|
|
# Store consolidated summary as a TEAM memory — only delete originals if POST succeeds
|
|
resp = await client.post(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories",
|
|
json={"content": f"[Consolidated] {summary}", "scope": "TEAM"},
|
|
headers=auth_headers(),
|
|
)
|
|
if resp.status_code in (200, 201):
|
|
# Safe to delete originals — consolidated version is saved
|
|
for m in memories:
|
|
await client.delete(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories/{m['id']}",
|
|
headers=auth_headers(),
|
|
)
|
|
logger.info("Consolidated %d memories into team knowledge", len(memories))
|
|
else:
|
|
logger.warning("Consolidation POST failed (status %d) — keeping originals", resp.status_code)
|
|
except Exception as e:
|
|
logger.error(
|
|
"CONSOLIDATION: Agent summarization failed (rate limit? model error?): %s. "
|
|
"Falling back to simple concatenation.", e
|
|
)
|
|
# Fall through to concatenation below
|
|
|
|
# Fallback: concatenate without agent summarization
|
|
if not (self.agent and summary):
|
|
combined = " | ".join(contents[:20])
|
|
await client.post(
|
|
f"{PLATFORM_URL}/workspaces/{WORKSPACE_ID}/memories",
|
|
json={"content": f"[Consolidated] {combined}", "scope": "TEAM"},
|
|
headers=auth_headers(),
|
|
)
|
|
logger.info("Consolidated %d memories via concatenation fallback", len(memories))
|
|
|
|
def stop(self):
|
|
self._running = False
|