- 📁 examples/
- 📄 .gitignore
- 📄 changelog.json
- 📄 CONTRIBUTING.md
Evidence-based endurance coaching protocol (v11.34). Use when analyzing training data, reviewing sessions, generating pre/post-workout reports, planning workouts, answering training questions, or giving endurance coaching advice. Always read or fetch athlete JSON data before responding to any training question.
- 📁 docs/
- 📁 examples/
- 📁 resources/
- 📄 SKILL.md
Complete CoinGecko Solana API integration for token prices, DEX pool data, OHLCV charts, trades, and market analytics. Use for building trading bots, portfolio trackers, price feeds, and on-chain data applications.
바선생 성장 리포트 — AI 활용 세션 데이터를 분석하여 성장 리포트를 자동 생성합니다. v2 레벨 시스템(6축×7단계, 0.5 단위)으로 분석합니다. "성장 리포트", "성장 분석", "얼마나 성장했는지", "레벨 체크", "성장 트래킹", "growth" 같은 요청에 사용됩니다.
- 📁 reference/
- 📄 README.md
- 📄 SKILL.md
Triggered when an "start awakening" related command is received. Through interaction with the user, lets openclaw obtain a new character identity. Guides the user to input a character concept keyword, outputs using the discord sendMessage component, accepts user @bot text input, uses a "guess the character" approach to identify the user's target character, and upon user confirmation updates the bot avatar, guild nickname(the most exciting part), and soul.md - transforming openclaw into that character.
从零构建机构级三表模型(IS/BS/CF)— 完整公式联动、季度/半年/年频自适应、IFRS/US GAAP/中国准则。 触发词:三表模型、financial model、3-statement、建模、从零建模、收入预测。 ❌ 填写已有模板请用 financial-analysis:3-statements --- # 3-Statement Model — IPO / Equity Research Quality (v4.8 · Public Edition) --- ## 🚀 Quick Start — New Users **This skill builds:** A complete institutional-grade 3-statement financial model (IS / BS / CF) in Excel, with full formula linkage, zero hardcoded forecast cells, and 9-step QC validation. **Prerequisites — install before starting:** ```bash pip install openpyxl yfinance pandas pip install notebooklm # optional — only needed if you have a NotebookLM notebook ``` **How to trigger:** Just say `"建个三表模型"` / `"build a 3-statement model for [Company]"` and the skill guides you step by step. **What to prepare:** - Company ticker (e.g. `BABA`, `0700.HK`, `600519.SS`) - A data source — see recommendations below - ~1–2 hours across 5 sessions (each session is independent — pause and resume anytime) **⚠️ Data Source Guide — Read Before Starting** | Option | Setup | Token Cost | Recommended? | |--------|-------|-----------|--------------| | **NotebookLM notebook** (annual reports / prospectus pre-loaded) | One-time OAuth auth setup | Very low — NLM handles the PDF; Claude only receives answers | ✅ Best path | | **Excel upload** (historical IS/BS/CF already structured) + short PDF excerpts | None | Low | ✅ Good | | **Direct PDF upload** (full annual report, prospectus) | None | 🔴 Very high — a single A-share annual report can be 200+ pages | ⚠️ Pro users: avoid | | **Web only** (Sina / Yahoo Finance fallback) | None | Low | ✅ Fallback | **Strongly recommended for new users: set up NotebookLM first.** The one-time auth flow takes ~5 minutes and saves significant token consumption for every future model: ```bash pip install notebooklm # Run once interactively — browser will open for Google OAuth python3 -c " import asyncio from notebooklm import NotebookLMClient async def auth(): async with await NotebookLMClient.
Structures AI and ML product decisions including model selection, data requirements, evaluation frameworks, and responsible AI considerations. Use when building AI-powered features, evaluating LLM integrations, designing AI products, or assessing AI readiness. Triggers on "AI product", "LLM feature", "AI canvas", "build with AI", "AI integration", "AI-powered", "machine learning feature".
Master advanced AgentDB features including QUIC synchronization, multi-database management, custom distance metrics, hybrid search, and distributed systems integration. Use when building distributed AI systems, multi-agent coordination, or advanced vector search applications.
Run a full static analysis of a project using spec-gen and summarise the results — architecture, call graph, top refactoring issues, and duplicate code. No LLM required.
SaaS churn reduction covering cancel flow design, dynamic save offers, exit survey architecture, dunning sequences, payment recovery, win-back campaigns, and churn impact modeling.
Debug AI traces, find exceptions, analyze sessions, and manage prompts via Langfuse MCP. Use when debugging AI pipelines, investigating errors, analyzing latency, managing prompt versions, or setting up Langfuse. Triggers on "langfuse", "traces", "debug AI", "find exceptions", "what went wrong", "why is it slow", "datasets", "evaluation sets".
Creates dbt models following project conventions. Use when working with dbt models for: (1) Creating new models (any layer - discovers project's naming conventions first) (2) Task mentions "create", "build", "add", "write", "new", or "implement" with model, table, or SQL (3) Modifying existing model logic, columns, joins, or transformations (4) Implementing a model from schema.yml specs or expected output requirements Discovers project conventions before writing. Runs dbt build (not just compile) to verify. --- # dbt Model Development **Read before you write. Build after you write. Verify your output.** ## Critical Rules 1. **ALWAYS run `dbt build`** after creating/modifying models - compile is NOT enough 2. **ALWAYS verify output** after build using `dbt show` - don't assume success 3. **If build fails 3+ times**, stop and reassess your entire approach ## Workflow ### 1. Understand the Task Requirements - What columns are needed? List them explicitly. - What is the grain of the table (one row per what)? - What calculations or aggregations are required? ### 2. Discover Project Conventions ```bash cat dbt_project.yml find models/ -name "*.sql" | head -20 ``` Read 2-3 existing models to learn naming, config, and SQL patterns. ### 3. Find Similar Models ```bash # Find models with similar purpose find models/ -name "*agg*.sql" -o -name "*fct_*.sql" | head -5 ```
Comprehensive analysis of BigQuery usage patterns, costs, and query performance