Use when building any system where email content triggers actions — AI agent inboxes, automated support handlers, email-to-task pipelines, or any workflow processing untrusted inbound email. Always use this skill when the user wants to receive emails and act on them programmatically, even if they don't mention "agent" — the skill contains critical security patterns (sender allowlists, content filtering, sandboxed processing) that prevent untrusted email from controlling your system.
This skill SHOULD be used when the user asks to "review code", "find dead code", "check for duplication", "simplify the codebase", "find refactoring opportunities", "do code cleanup", "check naming consistency", "analyze test organization", "run codebase health check", "review my PR", "refactor this code", "extract method", "rename variable", "consolidate duplicates", "adversarial review", "red team review", "find ways to break this", "multi-model review", "get multiple AI opinions on this code", "hunt bugs", "find bugs", "bug hunt", or "adversarial bug hunt". Routes to specialized analysis agents, refactoring workflow, multi-model adversarial review, or adversarial bug hunt based on the type of request.
Add "Open in molab" badge(s) linking to marimo notebooks. Works with READMEs, docs, websites, or any markdown/HTML target.
- 📁 evals/
- 📁 references/
- 📄 SKILL.md
Google Ads account audit and business context setup. Run this first — it gathers business information, analyzes account health, and saves context that all other ads skills reuse. Trigger on "audit my ads", "ads audit", "set up my ads", "onboard", "account overview", "how's my account", "ads health check", "what should I fix in my ads", or when the user is new to AdsAgent and hasn't run an audit before. Also trigger proactively when other ads skills detect that business-context.json is missing.
- 📁 examples/
- 📁 references/
- 📁 validation/
- 📄 README.md
- 📄 SKILL.md
- 📄 TROUBLESHOOTING.md
Use SwarmVault when the user needs a local-first knowledge vault that writes durable markdown, graph, search, dashboard, review, and MCP artifacts to disk from books, notes, transcripts, exports, datasets, slide decks, files, URLs, code, and recurring source workflows.
Admin and maintenance workflows for Ogham shared memory. Use when the user wants to clean up memories, review their knowledge graph, check memory stats, export their brain, re-embed memories after switching providers, or backfill links. Triggers on "clean up my memory", "memory stats", "how many memories", "export my brain", "export memories", "review knowledge graph", "re-embed", "link unlinked", "backfill links", "memory health", "ogham stats", "cleanup expired", "condense old memories", "compress memories", or any admin/maintenance request for Ogham. Requires the Ogham MCP server to be connected. --- # Ogham maintenance You handle admin tasks for Ogham shared memory. Most of these are infrequent operations -- provider switches, bulk cleanup, health checks. ## Available operations ### Health check Run `health_check` first if the user reports problems. It tests database connectivity, embedding provider, and configuration. Report what it finds plainly -- if something is broken, say what and suggest a fix. ### Stats overview Run `get_stats` and `list_profiles` to give the user a picture of their memory: - Total memories and breakdown by profile - Top sources (which clients are storing) - Top tags (what categories dominate) - Cache stats via `get_cache_stats` if they ask about performance Present it as a concise summary, not raw JSON. ### Cleanup expired memories 1. Run `get_stats` to show how many memories exist 2. Check if any profiles have TTLs set (this info comes from `list_profiles`) 3. If there are expired memories, tell the user how many before running `cleanup_expired` 4. Run `cleanup_expired` only after confirming with the user -- deletion is permanent ### Export Run `export_profile` with the format the user wants (JSON or Markdown). Tell them where the output goes and how to use it. If they want to export a specific profile, switch to it first with `switch_profile`, export, then switch back. ### Re-embed all memories This is needed after switching embedding
- 📁 assets/
- 📁 references/
- 📁 scripts/
- 📄 SKILL.md
Comprehensive EDA on scientific data files — structure, content, quality, and characteristics analysis across 200+ formats. Use when analyzing any data file to understand its structure, quality, and downstream analysis recommendations.
instruction layering with reusable, conditional instruction files
当用户希望把课堂逐字稿、OCR 笔记、截图资料或零散学习内容整理成结构化的 WPS 学习笔记时使用此 Skill。适合课堂记录整理、培训复盘、补课笔记汇总、课程主笔记构建等场景,尤其适合原始材料很乱、但最终需要一篇可复习、可回看、可继续补充的主笔记时。
- 📁 assets/
- 📁 config/
- 📁 craft/
- 📄 .gitignore
- 📄 novel_config.yaml
- 📄 pyproject.toml
Use when user wants to write novels, manage outlines, characters, world-building, style, workflows, or chapter review. Triggers: 写章节, 生成大纲, 创建角色, 世界观, 风格, 伏笔, 小说
Answer a question about a GRACE project using full project context. Use when the user has a question about the codebase, architecture, modules, or implementation — loads all GRACE artifacts, navigates the knowledge graph, and provides a grounded answer with citations.
- 📁 examples/
- 📄 ARCHITECTURE.md
- 📄 BENCHMARKS.md
- 📄 BUNDLES.md
Drop this file into your project (e.g. reference it from `CLAUDE.md`, `.cursor/rules`, or paste it into your AI context) so your AI assistant knows how to help you use Oxmgr.