docs: add docs/adapters/medo-smoke-test-log.md
This commit is contained in:
parent
891f36fe09
commit
96878a7c26
113
content/docs/adapters/medo-smoke-test-log.md
Normal file
113
content/docs/adapters/medo-smoke-test-log.md
Normal file
@ -0,0 +1,113 @@
|
||||
# MeDo Smoke Test Log — 2026-04-13 (Run 4)
|
||||
|
||||
**Tester:** PM (direct execution)
|
||||
**Goal:** Install Miaoda App Builder skill → build "Hello Molecule AI" landing page → publish → URL.
|
||||
**Credits spent:** 0 across all four runs.
|
||||
|
||||
---
|
||||
|
||||
## Run Summary
|
||||
|
||||
| Run | Blocker | Resolution |
|
||||
|-----|---------|------------|
|
||||
| 1 | `workspace-template:openclaw` image not built | ✅ Operator rebuilt image |
|
||||
| 2 | Adapter key lookup ignores `AISTUDIO_API_KEY` / `QIANFAN_API_KEY` | ✅ Code fix committed (d779e16) |
|
||||
| 3 | Executor creates fresh OpenClaw session per A2A message | ✅ Code fix committed (9466943) |
|
||||
| 4 | `payloads: []` on every response — agent never returns text via `--json` mode | ❌ Root cause below |
|
||||
|
||||
---
|
||||
|
||||
## Run 4 — Detailed Findings
|
||||
|
||||
### Environment — all green
|
||||
| Check | Result |
|
||||
|-------|--------|
|
||||
| Platform health | ✅ |
|
||||
| `workspace-template:openclaw` image | ✅ boots in 31s |
|
||||
| AISTUDIO_API_KEY + gemini-2.0-flash | ✅ confirmed in every response meta |
|
||||
| Stable session ID (workspace ID) | ✅ `sessionKey: agent:main:explicit:a507780d-...` consistent across all calls |
|
||||
|
||||
### Messages Sent and Responses
|
||||
|
||||
| Message | Response | Duration |
|
||||
|---------|----------|----------|
|
||||
| Install skill | `payloads: [], livenessState: working` | 1.7s |
|
||||
| Build Hello Molecule AI | `payloads: [], livenessState: working` | 0.8s |
|
||||
| Check status (sessions_list) | `LLM request failed: provider rejected request schema/payload` | — |
|
||||
| Reply with exactly: STATUS_OK | `payloads: [], livenessState: working` (after restart) | 1.8s |
|
||||
|
||||
The "Reply with exactly: STATUS_OK" response is decisive. A vanilla LLM call with no tool use should produce a text payload. It didn't. This rules out skill complexity or message ambiguity as the cause.
|
||||
|
||||
### Root Cause — `openclaw agent --json` Does Not Surface Agent Text in `payloads`
|
||||
|
||||
The OpenClaw agent processes messages using background session dispatch (`sessions_spawn` / `sessions_yield`). In this mode:
|
||||
1. Main session receives message → immediately spawns background session → calls `sessions_yield`
|
||||
2. `openclaw agent --json` exits with `payloads: [], livenessState: 'working'`
|
||||
3. Background session processes the actual work and produces text — but only visible in interactive/streaming mode, not in the `--json` subprocess call
|
||||
|
||||
**Evidence:** Even "Reply with exactly: STATUS_OK" returns `payloads: []`. The agent is using background sessions for everything, including trivial echo requests.
|
||||
|
||||
**Likely cause:** OpenClaw's default `SOUL.md` / `BOOTSTRAP.md` workspace config instructs the agent to always use async session patterns. In a terminal session these background responses appear naturally; via subprocess `--json`, only the main session's synchronous output is captured.
|
||||
|
||||
### Transient issue: LLM request failed
|
||||
After 3+ rapid A2A calls (install → build → status check), the Gemini AI Studio API returned a schema/payload rejection. Resolved by restarting the workspace (`POST /workspaces/:id/restart`). Likely a rate-limit or context-size rejection from Gemini. Restarted in 30s, normal on next call.
|
||||
|
||||
---
|
||||
|
||||
## 4. Required Fix — OpenClawA2AExecutor Response Capture
|
||||
|
||||
The executor must retrieve the agent's text response from session history **after** the main session yields. The `sessions_history` CLI command (exposed as `session_history` tool) retrieves past messages.
|
||||
|
||||
**Proposed change** to `workspace/adapters/openclaw/adapter.py` (`execute()` method):
|
||||
|
||||
```python
|
||||
# After proc.communicate() returns with payloads=[]:
|
||||
if not reply or reply.startswith("{'payloads': []"):
|
||||
# Agent yielded without responding — fetch last message from session history
|
||||
await asyncio.sleep(2) # brief wait for background session to complete short tasks
|
||||
hist_proc = await asyncio.create_subprocess_exec(
|
||||
"openclaw", "sessions", "history",
|
||||
"--session-id", self._session_id,
|
||||
"--limit", "1", "--json",
|
||||
stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE,
|
||||
env={**os.environ, "PATH": f"{os.path.expanduser('~/.local/bin')}:{os.environ.get('PATH', '')}"}
|
||||
)
|
||||
hist_stdout, _ = await asyncio.wait_for(hist_proc.communicate(), timeout=15)
|
||||
hist_data = json.loads(hist_stdout.decode().strip() or "{}")
|
||||
last_msg = (hist_data.get("messages") or [{}])[-1]
|
||||
reply = last_msg.get("content", reply) # fall back to original if no history
|
||||
```
|
||||
|
||||
**Note on long tasks (5–8 min builds):** Session history won't have the build result until it completes. For Miaoda App Builder, PM must poll: send a follow-up "What is the status of the Hello Molecule AI app build?" message every 60s until the response contains a URL or error.
|
||||
|
||||
---
|
||||
|
||||
## 5. Open Questions Status
|
||||
|
||||
### 5-C — Rate limits
|
||||
**UNKNOWN.** Never reached skill invocation.
|
||||
*New data:* Gemini AI Studio hit a schema/payload rejection after 3 rapid calls. This may be a Gemini-specific issue with large tool schemas (OpenClaw's `cron` schema is 6311 chars). Worth filing separately.
|
||||
|
||||
### 5-D — Failure recovery
|
||||
**UNKNOWN.** Never reached app generation.
|
||||
|
||||
---
|
||||
|
||||
## 6. Issues to File
|
||||
|
||||
| # | Issue | Status | Location |
|
||||
|---|-------|--------|----------|
|
||||
| A | `fix(openclaw): use stable workspace session ID` | ✅ fixed in 9466943 | adapter.py |
|
||||
| B | `fix(openclaw): extend key lookup for AISTUDIO/QIANFAN` | ✅ fixed in d779e16 | adapter.py |
|
||||
| C | `fix(provisioner): surface Docker errors in last_sample_error` | ❌ open | provisioner.go |
|
||||
| **D** | **`fix(openclaw): capture agent response via session history when payloads=[]`** | ❌ open — see §4 | adapter.py |
|
||||
| **E** | **`fix(openclaw): Gemini rejects request after N rapid calls with large tool schema`** | ❌ open — investigate cron schema size | adapter.py |
|
||||
|
||||
---
|
||||
|
||||
## 7. Next Steps (before Run 5)
|
||||
|
||||
- [ ] **Dev Lead:** Implement §4 session-history fallback in `OpenClawA2AExecutor.execute()`
|
||||
- [ ] **Dev Lead (optional):** Trim `cron` tool schema to reduce Gemini schema-size rejection risk
|
||||
- [ ] **Operator:** Rebuild image: `bash workspace/build-all.sh openclaw`
|
||||
- [ ] **PM (Run 5):** Re-run smoke test — expected to finally reach skill install confirmation
|
||||
Loading…
Reference in New Issue
Block a user