feat: EXP-15-19 training research, Gemma 4 adapter, data pipeline#377
Merged
CalebisGross merged 3 commits intomainfrom Apr 4, 2026
Merged
feat: EXP-15-19 training research, Gemma 4 adapter, data pipeline#377CalebisGross merged 3 commits intomainfrom
CalebisGross merged 3 commits intomainfrom
Conversation
…testing Research session covering rotation experiments, data scaling, Gemma 4 integration, and production quality validation. Experiments: - EXP-15/15b: orthogonal rotation (refuted full-space, minor bottleneck) - EXP-16: clean run 3 replication (eval 0.6074, 70% novel schema) - EXP-17: v2 dataset — 100% novel schema (removed poison data) - EXP-18: 12K encoding-only — 100% novel schema confirmed - EXP-19: Gemma 4 E2B + spokes — 100% schema, 5/7 stress test Infrastructure: - gemma_spoke_adapter.py (NF4, PLE offload, SpokeWrappedLayer) - batch_encode.py (Gemini Batch API, 50% cheaper) - compare_models.py, stress_test_hallucination.py - enrich_and_generate.py, extract_prenuke_data.py, merge_training_data.py - train_qwen_spokes.py --model-type gemma support - Handoff encoding preserved verbatim (agent.go) Results: Qwen 100% schema 20s/input vs Gemma 100% 34s/input vs Gemini 0%. Both spoke models 5/7 on hallucination stress test. Qwen selected as production encoding model for speed at equal quality. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
HTTP server on port 8899 that serves Qwen/Gemma + spokes as /v1/chat/completions endpoint. Allows mnemonic daemon to use the spoke model like LM Studio without GGUF export. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
Key additions
gemma_spoke_adapter.py— Gemma 4 E2B spoke adapter (NF4, PLE CPU offload, SpokeWrappedLayer)batch_encode.py— Gemini Batch API pipeline for scalable training data generationcompare_models.py— Side-by-side model comparison (schema, speed, quality)stress_test_hallucination.py— Hard input testing for detail preservationenrich_and_generate.py,extract_prenuke_data.py,merge_training_data.py— Data pipelinetrain_qwen_spokes.py— Updated with--model-type gemmasupport, OOM protectionexperiment_registry.md— Full results for EXP-15 through EXP-19agent.goTest plan
🤖 Generated with Claude Code