MetaClaw
MetaClaw is a continual learning proxy for OpenClaw and compatible agent runtimes — just talk to your agent and it learns from every conversation automatically.
About
MetaClaw is a continual learning proxy for OpenClaw and compatible agent runtimes — just talk to your agent and it learns from every conversation automatically. It sits behind an OpenAI-compatible proxy, injects relevant skills at each turn, and schedules RL-based meta-learning updates during idle windows so the agent improves without interrupting active use. No GPU cluster required; it uses Tinker or MinT for cloud-based LoRA training. Ranked #1 on HuggingFace Daily Papers, 134+ commits, actively developed.
ML engineers and research teams who want to experiment with continually adapting OpenClaw agents — particularly those exploring online fine-tuning, skill evolution, and learning from real user interactions rather than curated offline datasets.
Pros & Cons
Pros
- check No GPU required — cloud LoRA training via Tinker or MinT handles the heavy lifting
- check Skills are summarized and injected automatically after each session, providing immediate gains
- check Supports OpenClaw, IronClaw, PicoClaw, ZeroClaw, NanoClaw, NemoClaw, CoPaw, and OpenRouter
- check Hot-swap architecture means model updates don't interrupt live agent sessions
- check Academic-grade technical report, HuggingFace Daily Papers
Cons
- close Experimental research project — not production-hardened despite active development
- close RL training setup (Tinker/MinT backend) adds operational complexity and external dependencies
- close Privacy considerations: live conversations are captured as training data by design
- close MIT-licensed academic code may have rough edges in deployment scenarios
- close Learning from "the wild" means behavior can drift in unexpected directions if conversations are noisy
More Infrastructure
Other tools in the same category.