feat: add hippo-embed workflow + recurring embed step to daily-hippo-learn#28178
feat: add hippo-embed workflow + recurring embed step to daily-hippo-learn#28178
Conversation
- New hippo-embed.md: one-time workflow_dispatch maintenance workflow that audits (hippo audit --fix) and embeds (hippo embed) the Hippo memory store to restore semantic recall quality from < 1% coverage - Updated daily-hippo-learn.md: adds Step 2.5 to run hippo embed after every sleep cycle, keeping vector index current going forward - Compiled lock files regenerated for both workflows Agent-Logs-Url: https://github.com/github/gh-aw/sessions/ef7faf5b-5137-499d-adbd-ce805709fd6f Co-authored-by: pelikhan <4175913+pelikhan@users.noreply.github.com>
- Add runs-on: aw-gpu-runner-T4 (same runner as daily-news) - Move @xenova/transformers install into a frontmatter steps: block so it runs as a CI step before the agent, not via agent instruction - Remove the install section from the agent prompt Agent-Logs-Url: https://github.com/github/gh-aw/sessions/eaf7ade0-e041-429c-b704-16960fa6cbcc Co-authored-by: pelikhan <4175913+pelikhan@users.noreply.github.com>
|
Hey One thing worth addressing before this moves out of draft:
If you'd like a hand, here's a prompt to get started:
|
There was a problem hiding this comment.
Pull request overview
Adds a Hippo embedding maintenance workflow and introduces a recurring embedding step to keep the Hippo memory store’s vector index up to date for semantic recall.
Changes:
- Introduces a new
Hippo Embedworkflow (workflow_dispatchonly) that audits/prunes low-quality entries and embeds all memories. - Updates the daily Hippo learning workflow instructions to run
hippo embedafter eachhippo sleep. - Adds the new workflow to the agent-factory status documentation table.
Show a summary per file
| File | Description |
|---|---|
docs/src/content/docs/agent-factory-status.mdx |
Adds the Hippo Embed workflow to the published workflow/status list. |
.github/workflows/hippo-embed.md |
New maintenance workflow definition and agent instructions for audit + embed + status reporting. |
.github/workflows/hippo-embed.lock.yml |
Compiled/locked workflow artifact for hippo-embed.md. |
.github/workflows/daily-hippo-learn.md |
Adds an “embed” step after the daily “sleep” consolidation cycle. |
Copilot's findings
Tip
Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
- Files reviewed: 4/4 changed files
- Comments generated: 2
| ## Step 2.5 — Refresh embeddings | ||
|
|
||
| Keep the vector index current so semantic recall stays sharp. Run after every sleep | ||
| cycle to embed any memories that were added or updated since the last embed pass: | ||
|
|
||
| ``` | ||
| mcpscripts-hippo args: "embed" | ||
| ``` |
There was a problem hiding this comment.
The new mcpscripts-hippo args: "embed" step will fail in this workflow because @xenova/transformers (required by hippo embed, per the PR description) is never installed here. Add a setup steps: entry (similar to hippo-embed.md) to install the dependency, and re-run the workflow compiler so the change is reflected in daily-hippo-learn.lock.yml (the lock file is what GitHub Actions executes).
| steps: | ||
| - name: Install @xenova/transformers | ||
| run: | | ||
| npm install -g @xenova/transformers | ||
|
|
||
| imports: | ||
| - shared/hippo-memory.md |
There was a problem hiding this comment.
@xenova/transformers is installed here, but the same dependency is also needed by the recurring embed step added to daily-hippo-learn.md. To avoid duplicating setup across workflows (and to ensure future Hippo workflows don’t forget it), consider moving this install into a shared import (e.g., extend shared/hippo-memory.md or add a new shared hippo-embeddings.md import) and reusing it in both workflows.
The Hippo memory store has ~490 memories but fewer than 1% are vector-embedded, making semantic recall effectively non-functional. Root cause:
hippo embed(and its required dep@xenova/transformers) was never wired into any workflow.Changes
New
hippo-embed.md—workflow_dispatch-only maintenance workflow for the one-time fix, running onaw-gpu-runner-T4:hippo audit --fix— prunes flagged junk entries before indexingnpm install -g @xenova/transformers— installs the embedding backend as a CI step (before the agent runs)hippo embed— generates vectors for all memorieshippo status— reports coverage deltadaily-hippo-learn.md— adds Step 2.5 afterhippo sleepto runhippo embedon every daily cycle, keeping the vector index current as new memories accumulate:This is incremental — only unembedded memories are processed, so daily overhead is minimal.