- 📁 common_utils/
- 📁 find-testcases-covering-function/
- 📁 run-coverage/
- 📄 SKILL.md
Code coverage analysis tools. These tools help analyze and visualize code coverage for test execution, upload coverage data to Neo4j, and display coverage statistics. Available tools: run-coverage, show-coverage.
Run a full static analysis of a project using spec-gen and summarise the results — architecture, call graph, top refactoring issues, and duplicate code. No LLM required.
Bitbucket CLI for Data Center and Cloud. Use when users need to manage repositories, pull requests, branches, issues, webhooks, or pipelines in Bitbucket. Triggers include "bitbucket", "bkt", "pull request", "PR", "repo list", "branch create", "Bitbucket Data Center", "Bitbucket Cloud", "keyring timeout".
Comprehensive analysis of BigQuery usage patterns, costs, and query performance
Debug AI traces, find exceptions, analyze sessions, and manage prompts via Langfuse MCP. Use when debugging AI pipelines, investigating errors, analyzing latency, managing prompt versions, or setting up Langfuse. Triggers on "langfuse", "traces", "debug AI", "find exceptions", "what went wrong", "why is it slow", "datasets", "evaluation sets".
Ask a specific AI model (codex, gemini, grok, perplexity, claude) for focused analysis or a second opinion
Assay an experiment — deep analysis of results with cross-run comparison
Use the OfferPilot skill pack for resume optimization, China-first JD fit diagnosis, targeted resumes, and cover letters in Claude Code style repository agents. Use when the user wants structured job-application outputs or JD fit analysis from local resume and job-description files.
- 📁 examples/
- 📄 .gitignore
- 📄 changelog.json
- 📄 CONTRIBUTING.md
Evidence-based endurance coaching protocol (v11.25). Use when analyzing training data, reviewing sessions, generating pre/post-workout reports, planning workouts, answering training questions, or giving endurance coaching advice. Always read or fetch athlete JSON data before responding to any training question.
Design agent system prompts, parallel architectures, and methodological guardrails for data science decision-packs. Use when creating orchestrator, subagent, or parallel agent systems for analytical workflows. Covers anti-fabrication rules, epistemic humility, when to stop, conflict detection, uncertainty reporting, retry protocols, prompt design principles, and the decision-lab runtime mechanics.
Use for DataLion workflows such as listing, reading, creating, or editing projects, inspecting data sources, importing Excel or CSV data, working with reports and report tabs and codebooks, reading chart tables, or coordinating dashboard and export work through a configured datalion MCP server and related API or UI paths.
Build apps on Databricks Apps platform. Use when asked to create dashboards, data apps, analytics tools, or visualizations. Invoke BEFORE starting implementation.