Daily Featured Skills Count
3,840 3,909 3,920 3,927 3,966 4,007 4,027
04/06 04/07 04/08 04/09 04/10 04/11 04/12
♾️ Free & Open Source 🛡️ Secure & Worry-Free

Import Skills

contactjccoaching-wq contactjccoaching-wq
from GitHub Data & AI
  • 📁 agents/
  • 📁 benchmark/
  • 📁 benchmark-v4/
  • 📄 analysis.json
  • 📄 cheatsheet_memory.json
  • 📄 config.yaml

immune

Adaptive memory system that makes any LLM output better over time. Learns what works (strategies) and what fails (antibodies) from every scan. Injects winning patterns before generation, catches errors after. Hot/Cold tiered memory with multi-domain support.

0 10 7 days ago · Uploaded Detail →
geckse geckse
from GitHub Development & Coding
  • 📄 SKILL.md

prd-methodology

PRP (Product Requirement Prompt) methodology for writing PRDs. Reference for best practices in structuring requirements documents for coding agents. --- # PRP Methodology — Quick Reference The PRP (Product Requirement Prompt) framework is a structured process for creating PRDs that coding agents can execute in a single pass. ## Core Principle A PRD must contain ALL context needed for implementation. If a fresh Claude session with only the PRD can't build the feature correctly, the PRD is incomplete. ## The 3-Step Process 1. **Write initial description** — Brain dump what you want: feature, tech stack, constraints, integrations, examples, documentation references 2. **Generate the PRD** — Research the codebase + web, interview the user, produce a structured document following the base template 3. **Execute the PRD** — Clear context, start fresh, implement from the PRD alone ## What Makes a Good PRD **DO:** - Reference specific files and code patterns from the codebase - Write testable validation criteria ("returns 401 on invalid token") - Include explicit non-goals to prevent scope creep - List anti-patterns specific to the project - Order implementation steps by dependency (what must exist before what) - Include migration strategy for existing data/behavior **DON'T:** - Use vague validation criteria ("works well", "is performant") - Leave technical design abstract ("use appropriate data structures") - Assume the implementing agent knows project conventions — spell them out - Skip the non-goals section — agents will over-build without boundaries - Write steps that can't be verified independently ## Interview Technique The most valuable part of PRD generation is the interview. Goal: reduce assumptions to near zero. - Ask at least 8-10 questions before writing - Batch questions in groups of 3-4 - Provide recommended answers based on codebase research - Cover: scope, users, technical constraints, data model, compatibility, edge cases, testing, anti-patterns - Final ques

0 9 7 days ago · Uploaded Detail →
YehudaFrankel YehudaFrankel
from GitHub Content & Multimedia
  • 📄 SKILL.md

act

Reads recent session context, infers what you were working on, and proposes the specific next action. Use when resuming after a break, or say "act" / "what should we do next" / "pick up where we left off". Executes immediately on confirmation.

0 11 12 days ago · Uploaded Detail →
bytesquashcom bytesquashcom
from GitHub Tools & Productivity
  • 📄 SKILL.md

loggles-debug

Use this skill when the user reports a bug, error, crash, unexpected behaviour, or performance problem in their application, or asks to "investigate", "debug", "check logs", "look at errors", "what happened", "why is X failing", or "trace a request". Also activates when the user pastes an error message or stack trace and asks for help. Also use when the user asks "what is my app doing?", "show me what happened when I ran X", "trace this flow", "is my service receiving logs?", "I'm testing this endpoint — what do I see?", or any exploratory runtime question. Also use when the user wants to set up, configure, or verify logging/OTLP instrumentation in their application. Requires Loggles MCP tools to be connected.

0 10 12 days ago · Uploaded Detail →
ovlabs ovlabs
from GitHub Tools & Productivity
  • 📄 SKILL.md

brand-messaging-validation

Use this skill to validate and refine brand messaging, positioning, and taglines using real audience insight. Triggers include: requests to test brand positioning, validate taglines or slogans, understand brand perception, refine mission statements, test brand voice, or differentiate from competitors. Uses OriginalVoices Digital Twins (ask_twins) to understand how target audiences interpret brand messaging, what resonates emotionally, what creates confusion, and what drives brand affinity — ensuring brand statements connect with real people, not just internal stakeholders.

0 8 11 days ago · Uploaded Detail →
paulonasc paulonasc
from GitHub Tools & Productivity
  • 📁 bin/
  • 📁 changelog/
  • 📁 channels/
  • 📄 .gitignore
  • 📄 bun.lock
  • 📄 CONTRIBUTING.md

o

Multi-agent coordination through files. Memory, threads, progress, handoffs. Use when asked to "save progress", "checkpoint", "what's the status", "hand off", "what are we working on", or "close the thread".

0 7 8 days ago · Uploaded Detail →
ctoth ctoth
from GitHub Research & Analysis
  • 📄 SKILL.md

adjudicate

Systematically adjudicate disagreements across a paper collection. Produces ruthless verdicts on who was wrong, what supersedes what, and what the best current understanding is. Organized by topic clusters with actionable replacement values for implementation.

0 7 12 days ago · Uploaded Detail →
patrick-fu patrick-fu
from GitHub Development & Coding
  • 📁 evals/
  • 📁 references/
  • 📄 SKILL.md

brainstorm

Explore ideas, clarify goals, and help the user narrow down directions before planning or coding. Use this whenever the user proposes a new feature or idea, asks "what do you think about X", says "I'm thinking of building Y", wants to compare approaches, asks how to approach a problem, or seems to be exploring rather than ready to execute. Also use it when the user says "brainstorm", "let's think about this", "what's the best way to...", or any time the right next step is to clarify the problem and converge on a direction, not to write code yet. --- # Brainstorm Before anything else, ask. Don't jump to solutions or implementation. The goal is to draw out what the user actually means, uncover what they have not said yet, and help them converge on a direction. Think of this as Socratic dialogue with momentum: use questions to guide the thinking, but do not leave the user wandering in options forever. ## Start With Context Before asking, absorb the context that already exists in the conversation, codebase, docs, and project state. Do not ask for information you can already infer or look up directly. ## Guide The Conversation

0 5 7 days ago · Uploaded Detail →

Skill File Structure Sample (Reference)

skill-sample/
├─ SKILL.md              ⭐ Required: skill entry doc (purpose / usage / examples / deps)
├─ manifest.sample.json  ⭐ Recommended: machine-readable metadata (index / validation / autofill)
├─ LICENSE.sample        ⭐ Recommended: license & scope (open source / restriction / commercial)
├─ scripts/
│  └─ example-run.py     ✅ Runnable example script for quick verification
├─ assets/
│  ├─ example-formatting-guide.md  🧩 Output conventions: layout / structure / style
│  └─ example-template.tex         🧩 Templates: quickly generate standardized output
└─ references/           🧩 Knowledge base: methods / guides / best practices
   ├─ example-ref-structure.md     🧩 Structure reference
   ├─ example-ref-analysis.md      🧩 Analysis reference
   └─ example-ref-visuals.md       🧩 Visual reference

More Agent Skills specs Anthropic docs: https://agentskills.io/home

SKILL.md Requirements

├─ ⭐ Required: YAML Frontmatter (must be at top)
│  ├─ ⭐ name                 : unique skill name, follow naming convention
│  └─ ⭐ description          : include trigger keywords for matching
│
├─ ✅ Optional: Frontmatter extension fields
│  ├─ ✅ license              : license identifier
│  ├─ ✅ compatibility        : runtime constraints when needed
│  ├─ ✅ metadata             : key-value fields (author/version/source_url...)
│  └─ 🧩 allowed-tools        : tool whitelist (experimental)
│
└─ ✅ Recommended: Markdown body (progressive disclosure)
   ├─ ✅ Overview / Purpose
   ├─ ✅ When to use
   ├─ ✅ Step-by-step
   ├─ ✅ Inputs / Outputs
   ├─ ✅ Examples
   ├─ 🧩 Files & References
   ├─ 🧩 Edge cases
   ├─ 🧩 Troubleshooting
   └─ 🧩 Safety notes

Why SkillWink?

Skill files are scattered across GitHub and communities, difficult to search, and hard to evaluate. SkillWink organizes open-source skills into a searchable, filterable library you can directly download and use.

We provide keyword search, version updates, multi-metric ranking (downloads / likes / comments / updates), and open SKILL.md standards. You can also discuss usage and improvements on skill detail pages.

Keyword Search Version Updates Multi-Metric Ranking Open Standard Discussion

Quick Start:

Import/download skills (.zip/.skill), then place locally:

~/.claude/skills/ (Claude Code)

~/.codex/skills/ (Codex CLI)

One SKILL.md can be reused across tools.

FAQ

Everything you need to know: what skills are, how they work, how to find/import them, and how to contribute.

1. What are Agent Skills?

A skill is a reusable capability package, usually including SKILL.md (purpose/IO/how-to) and optional scripts/templates/examples.

Think of it as a plugin playbook + resource bundle for AI assistants/toolchains.

2. How do Skills work?

Skills use progressive disclosure: load brief metadata first, load full docs only when needed, then execute by guidance.

This keeps agents lightweight while preserving enough context for complex tasks.

3. How can I quickly find the right skill?

Use these three together:

  • Semantic search: describe your goal in natural language.
  • Multi-filtering: category/tag/author/language/license.
  • Sort by downloads/likes/comments/updated to find higher-quality skills.

4. Which import methods are supported?

  • Upload archive: .zip / .skill (recommended)
  • Upload skills folder
  • Import from GitHub repository

Note: file size for all methods should be within 10MB.

5. How to use in Claude / Codex?

Typical paths (may vary by local setup):

  • Claude Code:~/.claude/skills/
  • Codex CLI:~/.codex/skills/

One SKILL.md can usually be reused across tools.

6. Can one skill be shared across tools?

Yes. Most skills are standardized docs + assets, so they can be reused where format is supported.

Example: retrieval + writing + automation scripts as one workflow.

7. Are these skills safe to use?

Some skills come from public GitHub repositories and some are uploaded by SkillWink creators. Always review code before installing and own your security decisions.

8. Why does it not work after import?

Most common reasons:

  • Wrong folder path or nested one level too deep
  • Invalid/incomplete SKILL.md fields or format
  • Dependencies missing (Python/Node/CLI)
  • Tool has not reloaded skills yet

9. Does SkillWink include duplicates/low-quality skills?

We try to avoid that. Use ranking + comments to surface better skills:

  • Duplicate skills: compare differences (speed/stability/focus)
  • Low quality skills: regularly cleaned up