Daily Featured Skills Count
3,626 3,840 3,909 3,920 3,927 3,966 4,007
04/05 04/06 04/07 04/08 04/09 04/10 04/11
♾️ Free & Open Source 🛡️ Secure & Worry-Free

Import Skills

sun-guannan sun-guannan
from GitHub Content & Multimedia
  • 📁 references/
  • 📁 scripts/
  • 📄 SKILL.md

vectcut-api

VectCutAPI is a powerful cloud-based video editing API tool that provides programmatic control over CapCut/JianYing (剪映) for professional video editing. Use this skill when users need to: (1) Create video draft projects programmatically, (2) Add video/audio/image materials with precise control, (3) Add text, subtitles, and captions, (4) Apply effects, transitions, and animations, (5) Add keyframe animations, (6) Process videos in batch, (7) Generate AI-powered videos, (8) Integrate with n8n workflows, (9) Build MCP video editing agents. The API supports HTTP REST and MCP protocols, works with both CapCut (international) and JianYing (China), and provides web preview without downloading.

0 1.8K 9 days ago · Uploaded Detail →
daymade daymade
from GitHub Content & Multimedia
  • 📁 references/
  • 📁 scripts/
  • 📄 .security-scan-passed
  • 📄 SKILL.md

asr-transcribe-to-text

Transcribes audio and video files to text using Qwen3-ASR. Supports two modes — local MLX inference on macOS Apple Silicon (no API key, 15-27x realtime) and remote API via vLLM/OpenAI-compatible endpoints. Auto-detects platform and recommends the best path. Triggers when the user wants to transcribe recordings, convert audio/video to text, do speech-to-text, or mentions ASR, Qwen ASR, 转录, 语音转文字, 录音转文字. Also triggers for meeting recordings, lectures, interviews, podcasts, screen recordings, or any audio/video file the user wants converted to text.

0 782 8 days ago · Uploaded Detail →
heygen-com heygen-com
from GitHub Content & Multimedia
  • 📄 SKILL.md

ai-video-gen

Generate AI videos from text prompts using the HeyGen API. Use when: (1) Generating videos from text descriptions, (2) Creating AI-generated video clips for content production, (3) Image-to-video generation with a reference image, (4) Choosing between video generation providers (VEO, Kling, Sora, Runway, Seedance), (5) Working with HeyGen's /v1/workflows/executions endpoint for video generation.

0 102 5 days ago · Uploaded Detail →
cerul-ai cerul-ai
from GitHub Content & Multimedia
  • 📄 SKILL.md

cerul

You cannot access video content on your own. Use Cerul to search what was said, shown, or presented in tech talks, podcasts, conference presentations, and earnings calls. Use when a user asks about what someone said, wants video evidence, or needs citations from talks and interviews.

0 92 7 days ago · Uploaded Detail →
BintzGavin BintzGavin
from GitHub Tools & Productivity
  • 📄 SKILL.md

helios-skills

Collection of agent skills for Helios video engine. Use when working with programmatic video creation, browser-native animations, or Helios compositions. Install individual skills by path for specific capabilities.

0 61 6 days ago · Uploaded Detail →
JimmyLv JimmyLv
from GitHub Content & Multimedia
  • 📁 references/
  • 📁 scripts/
  • 📁 workflows/
  • 📄 SKILL.md

bibi

AI video & audio summarizer. Summarize YouTube videos, Bilibili videos, podcasts, TikTok, Twitter/X, Xiaohongshu, and any online video or audio. Use when the user wants to summarize a video, extract transcripts/subtitles, get chapter-by-chapter summaries, or understand video content quickly.

0 38 8 days ago · Uploaded Detail →
woodfantasy woodfantasy
from GitHub Content & Multimedia
  • 📁 references/
  • 📁 scripts/
  • 📄 .gitignore
  • 📄 LICENSE
  • 📄 README.es.md

seedance-shot-design

Professional-grade virtual film director and prompt engineer for Seedance 2.0 (即梦). Transforms vague ideas into cinematic, production-ready video prompts with Hollywood-caliber shot design. Covers every workflow — text-to-video, image-to-video, multi-modal references, video extension, character swap, dialogue-driven short films, and music-synced edits. Ships with a cinematography dictionary (50+ safe camera-move phrases), a director style library (Villeneuve, Wes Anderson, Shinkai, Wuxia & more), a 3-layer lighting & quality-anchor system that kills the "plastic AI look," and built-in Python auto-validation so every prompt passes before delivery. Supports bilingual output (Chinese/English) with smart >15 s auto-segmentation for long-form storytelling. Trigger words: Seedance, Shot Design, AI video, storyboard, video prompt, short film, ad video, film prompt, cinematic prompt, generate a video, make a clip, shoot a scene, video script, vlog script, create video prompt, music video, product video, drone shot, camera movement, 即梦, 视频提示词, 分镜, 帮我写个视频, 帮我拍, 做个视频脚本, 写一段视频, 生成视频, 视频文案, 短视频, 拍一个, 做分镜, 视频脚本, AI视频, 抖音视频, 短片脚本, 广告视频, 宣传片, 产品视频, vlog, 运镜, 镜头设计.

0 27 12 days ago · Uploaded Detail →
swyxio swyxio
from GitHub Content & Multimedia
  • 📄 README.md
  • 📄 SKILL.md

download-video

Downloads embedded videos from web pages. Fetches the page, identifies the video hosting service (Vimeo, YouTube, etc.), resolves the correct embed/player URL, and downloads using yt-dlp. Handles private/unlisted videos that require referer headers or embed URLs. Use this skill when someone says "download this video", "save this video", "grab the video from this page", "rip this video", or provides a URL and asks to download media from it. Also trigger when someone pastes a URL to a page with an embedded video and wants the video file locally.

0 21 10 days ago · Uploaded Detail →
NVIDIA-AI-IOT NVIDIA-AI-IOT
from GitHub Development & Coding
  • 📁 references/
  • 📄 SKILL.md

deepstream-dev

NVIDIA DeepStream SDK 9.0 development with Python pyservicemaker API. Use when building video analytics pipelines, GStreamer-based video processing, TensorRT inference integration, object detection/tracking, or Kafka/message broker integration.

0 20 11 days ago · Uploaded Detail →
maxazure maxazure
from GitHub Content & Multimedia
  • 📁 docs/
  • 📁 remotion-standup/
  • 📁 scripts/
  • 📄 .gitignore
  • 📄 README.md
  • 📄 REMOTION_VOICEOVER.md

video-editing

Automated video editing skill for talk/vlog/standup videos. Use when: cutting video, splitting video into sentences, merging video clips, extracting audio, transcribing speech, auto-editing oral presentation videos, combining selected sentence clips into a final video, generating video cover/thumbnail with title, B-roll cutaway editing, persistent video overlay/watermark, blinking REC indicator, ending title cards, multi-source audio mixing, generating voiceover videos with Remotion (audio-only to video with animated visuals/subtitles). Requires ffmpeg and whisper. Remotion workflow additionally requires Node.js and npm.

0 14 10 days ago · Uploaded Detail →
chencore chencore
from GitHub Content & Multimedia
  • 📄 skill.md

seedance-creator

This skill should be used when the user asks to "generate video prompts", "create Seedance prompts", "write video descriptions", mentions "Seedance", "seedance", "即梦", "即梦平台", "视频提示词", "视频生成", "AI视频", "短剧", "广告视频", "视频延长", "生成图片", "文生图", "图生图", "图生视频", "文生视频", or discusses video prompt engineering, AI video generation, or Seedance 2.0 workflows. It also handles requests to create, edit, or manipulate images and videos using the dreamina CLI tool.

0 8 6 days ago · Uploaded Detail →

Skill File Structure Sample (Reference)

skill-sample/
├─ SKILL.md              ⭐ Required: skill entry doc (purpose / usage / examples / deps)
├─ manifest.sample.json  ⭐ Recommended: machine-readable metadata (index / validation / autofill)
├─ LICENSE.sample        ⭐ Recommended: license & scope (open source / restriction / commercial)
├─ scripts/
│  └─ example-run.py     ✅ Runnable example script for quick verification
├─ assets/
│  ├─ example-formatting-guide.md  🧩 Output conventions: layout / structure / style
│  └─ example-template.tex         🧩 Templates: quickly generate standardized output
└─ references/           🧩 Knowledge base: methods / guides / best practices
   ├─ example-ref-structure.md     🧩 Structure reference
   ├─ example-ref-analysis.md      🧩 Analysis reference
   └─ example-ref-visuals.md       🧩 Visual reference

More Agent Skills specs Anthropic docs: https://agentskills.io/home

SKILL.md Requirements

├─ ⭐ Required: YAML Frontmatter (must be at top)
│  ├─ ⭐ name                 : unique skill name, follow naming convention
│  └─ ⭐ description          : include trigger keywords for matching
│
├─ ✅ Optional: Frontmatter extension fields
│  ├─ ✅ license              : license identifier
│  ├─ ✅ compatibility        : runtime constraints when needed
│  ├─ ✅ metadata             : key-value fields (author/version/source_url...)
│  └─ 🧩 allowed-tools        : tool whitelist (experimental)
│
└─ ✅ Recommended: Markdown body (progressive disclosure)
   ├─ ✅ Overview / Purpose
   ├─ ✅ When to use
   ├─ ✅ Step-by-step
   ├─ ✅ Inputs / Outputs
   ├─ ✅ Examples
   ├─ 🧩 Files & References
   ├─ 🧩 Edge cases
   ├─ 🧩 Troubleshooting
   └─ 🧩 Safety notes

Why SkillWink?

Skill files are scattered across GitHub and communities, difficult to search, and hard to evaluate. SkillWink organizes open-source skills into a searchable, filterable library you can directly download and use.

We provide keyword search, version updates, multi-metric ranking (downloads / likes / comments / updates), and open SKILL.md standards. You can also discuss usage and improvements on skill detail pages.

Keyword Search Version Updates Multi-Metric Ranking Open Standard Discussion

Quick Start:

Import/download skills (.zip/.skill), then place locally:

~/.claude/skills/ (Claude Code)

~/.codex/skills/ (Codex CLI)

One SKILL.md can be reused across tools.

FAQ

Everything you need to know: what skills are, how they work, how to find/import them, and how to contribute.

1. What are Agent Skills?

A skill is a reusable capability package, usually including SKILL.md (purpose/IO/how-to) and optional scripts/templates/examples.

Think of it as a plugin playbook + resource bundle for AI assistants/toolchains.

2. How do Skills work?

Skills use progressive disclosure: load brief metadata first, load full docs only when needed, then execute by guidance.

This keeps agents lightweight while preserving enough context for complex tasks.

3. How can I quickly find the right skill?

Use these three together:

  • Semantic search: describe your goal in natural language.
  • Multi-filtering: category/tag/author/language/license.
  • Sort by downloads/likes/comments/updated to find higher-quality skills.

4. Which import methods are supported?

  • Upload archive: .zip / .skill (recommended)
  • Upload skills folder
  • Import from GitHub repository

Note: file size for all methods should be within 10MB.

5. How to use in Claude / Codex?

Typical paths (may vary by local setup):

  • Claude Code:~/.claude/skills/
  • Codex CLI:~/.codex/skills/

One SKILL.md can usually be reused across tools.

6. Can one skill be shared across tools?

Yes. Most skills are standardized docs + assets, so they can be reused where format is supported.

Example: retrieval + writing + automation scripts as one workflow.

7. Are these skills safe to use?

Some skills come from public GitHub repositories and some are uploaded by SkillWink creators. Always review code before installing and own your security decisions.

8. Why does it not work after import?

Most common reasons:

  • Wrong folder path or nested one level too deep
  • Invalid/incomplete SKILL.md fields or format
  • Dependencies missing (Python/Node/CLI)
  • Tool has not reloaded skills yet

9. Does SkillWink include duplicates/low-quality skills?

We try to avoid that. Use ranking + comments to surface better skills:

  • Duplicate skills: compare differences (speed/stability/focus)
  • Low quality skills: regularly cleaned up