Daily Featured Skills Count
3,840 3,909 3,920 3,927 3,966 4,007 4,027
04/06 04/07 04/08 04/09 04/10 04/11 04/12
♾️ Free & Open Source 🛡️ Secure & Worry-Free

Import Skills

Okayamasenko Okayamasenko
from GitHub Research & Analysis
  • 📁 commands/
  • 📁 memory/
  • 📁 notes/
  • 📄 .gitignore
  • 📄 CLAUDE.md
  • 📄 config.json

articlefeed

Academic paper reading system — find papers via Semantic Scholar, read with structured notes, discuss with auto-recording, track progress. 7 commands: /setup /feed /read /discuss /recap /update /sync

0 17 6 days ago · Uploaded Detail →
lunchpaillola lunchpaillola
from GitHub Research & Analysis
  • 📁 references/
  • 📄 SKILL.md

autoresearch

Autonomously optimize any Claude Code skill by running it repeatedly, scoring outputs against binary evals, mutating the prompt, and keeping improvements. Based on Karpathy's autoresearch methodology. Use when: optimize this skill, improve this skill, run autoresearch on, make this skill better, self-improve skill, benchmark skill, eval my skill, run evals on. Outputs: an improved SKILL.md, a results log, and a changelog of every mutation tried.

0 20 10 days ago · Uploaded Detail →
punt-labs punt-labs
from GitHub Research & Analysis
  • 📁 references/
  • 📄 SKILL.md

prfaq

This skill should be used when the user asks to "write a PR/FAQ", "prfaq", "working backwards", "product discovery", "evaluate a product idea", "press release FAQ", "test product value", "revise prfaq", "update prfaq", "add research to prfaq", "add FAQs", "run a meeting", "review meeting", "hive meeting", "autonomous meeting", "consensus meeting", "stress test my prfaq", "go/no-go decision", "should we build this", "vote on prfaq", or wants to use the Amazon Working Backwards process to evaluate whether a product or feature is worth building. --- # Working Backwards: PR/FAQ ## Purpose Guide the user through the Amazon Working Backwards process to produce a professional PR/FAQ document. The output is a LaTeX file that compiles to a polished PDF suitable for executive review and product decision-making. The process forces clarity about customer value, surfaces risks early, and creates a shared artifact for go/no-go decisions. ## When to Use - Evaluating whether a new product or feature is worth building - Forcing specificity on a vague product idea - Preparing a product pitch for leadership review - Testing whether a team truly understands the customer problem - Structuring a go/no-go decision with an auditable artifact ## Revise Mode Before starting the full workflow, check if a `prfaq.tex` file already exists in the project root (or the path the user specifies). If it does, enter **revise mode** instead of starting from scratch. 1. **Read the existing document.** Parse the `.tex` file to understand what's already written — the press release, FAQs, and risk assessment. 2. **Ask what to revise.** Present the user with the sections found and ask what they want to improve. Common revision goals: - **Refine the product** — sharpen the problem statement, solution, or differentiation based on new thinking - **Incorporate research** — thread new primary data (customer interviews, market analysis, survey results) into existing sections. Run Phase 0 research discovery to find

0 17 12 days ago · Uploaded Detail →
Pthahnix Pthahnix
from GitHub Research & Analysis
  • 📁 references/
  • 📄 SKILL.md

DARE

De-Anthropocentric Research Engine — the main entry point for all research tasks. Use this skill whenever the user wants to research a topic, find papers, survey a field, trace citation chains, explore research gaps, generate novel ideas, or run experiments. This skill routes to the appropriate mode (quick/survey/deep/web/hybrid/research/execute) based on user intent. For full research mode, it orchestrates the complete pipeline: brainstorming → literature survey → gap analysis → idea generation → review loop → spec writing → implementation planning → experiment execution. ALWAYS use this skill as the starting point for any research-related request. --- # DARE DARE is a Vibe Researching Toolkit. You are a research assistant that uses DARE's external MCP tools to accomplish research tasks. ## Your Role You are an autonomous research agent. Given a research topic or question, you: 1. Understand intent, gauge desired depth and breadth 2. Automatically select the appropriate mode 3. Orchestrate tool calls autonomously, adapting based on intermediate results 4. Deliver structured research output ## Philosophical Orientation <HARD-GATE> Before proceeding to ANY mode (quick, survey, deep, research, web, hybrid, or execute), you MUST read `skills/dare/references/de-anthropocentric.md` in full. This is the value system that governs all research activity in DARE. Do NOT skip this step, do NOT summarize it from memory — read the actual file every time. </HARD-GATE> This philosophy is not a set of instructions. It is a lens. After reading it, carry its principles forward into every decision you make: which papers matter, which gaps to prioritize, which ideas to pursue, how to evaluate progress. The document will tell you what those principles are. ## Tools See `skills/tools/SKILL.md` for full reference, `skills/tools/references/dare-scholar.md` for detailed paper tool usage. | MCP Server | Tool | Purpose | |---|---|---| | apify | `google-scholar-scraper` | Google Scholar search

0 12 6 days ago · Uploaded Detail →
comet-ml comet-ml
from GitHub Research & Analysis
  • 📁 references/
  • 📄 SKILL.md

agent-ops

This skill should be used when the user asks about agent architecture, evaluation, metrics, production monitoring, debugging agents, or best practices for building reliable AI agents. Use for questions like "evaluate my agent", "set up production monitoring", "add guardrails", "detect hallucinations", "agent anti-patterns", "compare experiments", "create evaluation dataset".

0 14 11 days ago · Uploaded Detail →
viitrix viitrix
from GitHub Research & Analysis
  • 📄 SKILL.md

agent-browser

Browse the web for any task — research topics, read articles, interact with web apps, fill forms, take screenshots, extract data, and test web pages. Use whenever a browser would be useful, not just when the user explicitly asks.

0 14 12 days ago · Uploaded Detail →
PostHog PostHog
from GitHub Research & Analysis
  • 📄 SKILL.md

analyzing-experiment-session-replays

Analyze session replay patterns across experiment variants to understand user behavior differences. Use when the user wants to see how users interact with different experiment variants, identify usability issues, compare behavior patterns between control and test groups, or get qualitative insights to complement quantitative experiment results.

0 13 10 days ago · Uploaded Detail →

Skill File Structure Sample (Reference)

skill-sample/
├─ SKILL.md              ⭐ Required: skill entry doc (purpose / usage / examples / deps)
├─ manifest.sample.json  ⭐ Recommended: machine-readable metadata (index / validation / autofill)
├─ LICENSE.sample        ⭐ Recommended: license & scope (open source / restriction / commercial)
├─ scripts/
│  └─ example-run.py     ✅ Runnable example script for quick verification
├─ assets/
│  ├─ example-formatting-guide.md  🧩 Output conventions: layout / structure / style
│  └─ example-template.tex         🧩 Templates: quickly generate standardized output
└─ references/           🧩 Knowledge base: methods / guides / best practices
   ├─ example-ref-structure.md     🧩 Structure reference
   ├─ example-ref-analysis.md      🧩 Analysis reference
   └─ example-ref-visuals.md       🧩 Visual reference

More Agent Skills specs Anthropic docs: https://agentskills.io/home

SKILL.md Requirements

├─ ⭐ Required: YAML Frontmatter (must be at top)
│  ├─ ⭐ name                 : unique skill name, follow naming convention
│  └─ ⭐ description          : include trigger keywords for matching
│
├─ ✅ Optional: Frontmatter extension fields
│  ├─ ✅ license              : license identifier
│  ├─ ✅ compatibility        : runtime constraints when needed
│  ├─ ✅ metadata             : key-value fields (author/version/source_url...)
│  └─ 🧩 allowed-tools        : tool whitelist (experimental)
│
└─ ✅ Recommended: Markdown body (progressive disclosure)
   ├─ ✅ Overview / Purpose
   ├─ ✅ When to use
   ├─ ✅ Step-by-step
   ├─ ✅ Inputs / Outputs
   ├─ ✅ Examples
   ├─ 🧩 Files & References
   ├─ 🧩 Edge cases
   ├─ 🧩 Troubleshooting
   └─ 🧩 Safety notes

Why SkillWink?

Skill files are scattered across GitHub and communities, difficult to search, and hard to evaluate. SkillWink organizes open-source skills into a searchable, filterable library you can directly download and use.

We provide keyword search, version updates, multi-metric ranking (downloads / likes / comments / updates), and open SKILL.md standards. You can also discuss usage and improvements on skill detail pages.

Keyword Search Version Updates Multi-Metric Ranking Open Standard Discussion

Quick Start:

Import/download skills (.zip/.skill), then place locally:

~/.claude/skills/ (Claude Code)

~/.codex/skills/ (Codex CLI)

One SKILL.md can be reused across tools.

FAQ

Everything you need to know: what skills are, how they work, how to find/import them, and how to contribute.

1. What are Agent Skills?

A skill is a reusable capability package, usually including SKILL.md (purpose/IO/how-to) and optional scripts/templates/examples.

Think of it as a plugin playbook + resource bundle for AI assistants/toolchains.

2. How do Skills work?

Skills use progressive disclosure: load brief metadata first, load full docs only when needed, then execute by guidance.

This keeps agents lightweight while preserving enough context for complex tasks.

3. How can I quickly find the right skill?

Use these three together:

  • Semantic search: describe your goal in natural language.
  • Multi-filtering: category/tag/author/language/license.
  • Sort by downloads/likes/comments/updated to find higher-quality skills.

4. Which import methods are supported?

  • Upload archive: .zip / .skill (recommended)
  • Upload skills folder
  • Import from GitHub repository

Note: file size for all methods should be within 10MB.

5. How to use in Claude / Codex?

Typical paths (may vary by local setup):

  • Claude Code:~/.claude/skills/
  • Codex CLI:~/.codex/skills/

One SKILL.md can usually be reused across tools.

6. Can one skill be shared across tools?

Yes. Most skills are standardized docs + assets, so they can be reused where format is supported.

Example: retrieval + writing + automation scripts as one workflow.

7. Are these skills safe to use?

Some skills come from public GitHub repositories and some are uploaded by SkillWink creators. Always review code before installing and own your security decisions.

8. Why does it not work after import?

Most common reasons:

  • Wrong folder path or nested one level too deep
  • Invalid/incomplete SKILL.md fields or format
  • Dependencies missing (Python/Node/CLI)
  • Tool has not reloaded skills yet

9. Does SkillWink include duplicates/low-quality skills?

We try to avoid that. Use ranking + comments to surface better skills:

  • Duplicate skills: compare differences (speed/stability/focus)
  • Low quality skills: regularly cleaned up