- 📁 references/
- 📁 tests/
- 📄 SKILL.md
Use when implementing any feature, adding code, or modifying existing code in this Kotlin/Spring project. Triggers on write operations like adding entities, services, facades, controllers, or any domain logic.
- 📁 assets/
- 📁 reference/
- 📄 SKILL.md
MoonBit code generation best practices. Use when writing MoonBit code to avoid common AI mistakes with syntax, tests, and benchmarks.
Run a code quality review on changed files after code modifications. Triggers when application logic, architecture, data flow, or reusable components are affected.
Use this skill for any request to create, update, review, or improve files that guide AI coding tools—like AGENTS.md, Copilot context files, or agent instructions. Trigger when users want to help AI generate code that matches project conventions, avoids common mistakes, or understands non-obvious rules—whether for a whole repo, a subdirectory, or a specific component. Also use for queries about setting up Copilot or Cursor context, onboarding AI to team practices, or keeping agent guidance up to date—even if AGENTS.md isn't mentioned by name. --- # AGENTS.md Skill `AGENTS.md` is the README for AI coding agents — it gives tools like GitHub Copilot, Cursor, and Claude Code the context they need to produce code that fits your project, without constant back-and-forth correction. ## The Golden Rule: Only Include Non-Obvious Things This is the most important principle. Before adding any line, ask: **"Could an agent figure this out by reading the code or config files?"** If yes — skip it. Agents can read `package.json`, `pyproject.toml`, `*.csproj`, directory structures, imports, and existing code. AGENTS.md is for the things that *aren't* visible there: - Project-specific conventions not enforced by any linter or analyzer - "Never do X" patterns that *look* reasonable but are wrong in this codebase - Commands that are non-standard or require project-specific flags - Domain terminology or architectural choices that need explanation - Common mistakes agents actually make here (discovered from experience) - Security constraints and rules that must never be violated **Redundant content actively degrades quality** — it wastes the agent's context window and dilutes real signal with noise it already has. ## Two Shapes of AGENTS.md ### Root-level `AGENTS.md` (repo root)
or non-idiomatic Go code in their project. Also when asked "how good is this code?" or "audit my code."
- 📁 .claude-plugin/
- 📁 references/
- 📄 SKILL.md
Codebase analyzer for JavaScript/TypeScript projects. Finds unused code (files, exports, types, dependencies), code duplication, circular dependencies, complexity hotspots, architecture boundary violations, and feature flag patterns. 85 framework plugins, zero configuration, sub-second performance. Use when asked to analyze code health, find unused code, detect duplicates, check circular dependencies, audit complexity, check architecture boundaries, detect feature flags, clean up the codebase, auto-fix issues, or run fallow.
High-confidence code security review workflow for changed code, using modern threat-informed methodologies with strict false-positive filtering and exploit-focused findings.
Audit web applications and codebases for the most common and dangerous security vulnerabilities — especially those introduced by AI-assisted ("vibe coded") development. Use this skill whenever the user asks to review code for security issues, harden an app, audit an API, check for vulnerabilities, or secure a project. Also trigger when the user mentions terms like "security review", "pentest checklist", "harden my app", "is my code secure", "fix security holes", "OWASP", "SQL injection", "XSS", "vibe code security", or shares backend/frontend code and asks if anything looks wrong. Even if the user just says "review my code" without mentioning security, consider triggering this skill — security is always relevant. --- # Vibe-Code Security Audit Systematic security audit for web applications, with special attention to vulnerabilities that AI code-generation tools introduce most frequently.
Find functions with high cyclomatic complexity, excessive length, or too many parameters. Use when the user asks to find complex code, complexity hotspots, refactoring candidates, or wants to improve code maintainability.
This skill should be used when the user asks to "calculate code mass", "measure code complexity with APP", "compare implementations using APP", "apply Absolute Priority Premise", "use mass calculations", or during TDD refactor phases when comparing alternative implementations. Based on Micah Martin's work.
Autonomous iterative experimentation loop for any programming task. Guides the user through defining goals, measurable metrics, and scope constraints, then runs an autonomous loop of code changes, testing, measuring, and keeping/discarding results. Inspired by Karpathy''s autoresearch. USE FOR: autonomous improvement, iterative optimization, experiment loop, auto research, performance tuning, automated experimentation, hill climbing, try things automatically, optimize code, run experiments, autonomous coding loop. DO NOT USE FOR: one-shot tasks, simple bug fixes, code review, or tasks without a measurable metric.
Pre-commit code quality review for Unity C# code. Checks against Rules, CODING_STANDARDS.md, NAMING_CONVENTIONS.md, and GDD Gherkin test coverage. Use when code has been written and tests are passing but before committing. Runs automatically at Step 8.5 of /uw-cmd-implement-feature. Triggers on "review this code", "check my implementation", "pre-commit review", "code review", "check before commit", "is this ready to commit", "review my changes", or any request to validate code quality before committing.