Anthropic Skills
2689 skills. Last updated 2026-03-15
Discover and use Anthropic Skills to extend Claude's capabilities with creative, technical, and enterprise workflows.
Generate draw.io editable diagrams (.drawio, .drawio.svg) from text, images, or Excel. Orchestrates 3-agent workflow (Analysis → Manifest → SVG generation) with quality gates. Use when creating architecture diagrams, flowcharts, sequence diagrams, or converting existing images to editable format. Supports Azure/AWS cloud icons.
WCAG-compliant Mermaid diagrams using verified accessible color palette. Use when creating diagrams, flowcharts, or any color-dependent visualizations requiring accessibility compliance for color blindness.
Explores data in a Bauplan lakehouse safely using the Bauplan Python SDK. Use to inspect namespaces, tables, schemas, samples, and profiling queries; and to export larger result sets to files. Read-only exploration only; no writes or pipeline runs.
Expert guidance for Mermaid.js diagramming library. Create flowcharts, sequence diagrams, class diagrams, state diagrams, Gantt charts, git graphs, and block diagrams. Use when working with Mermaid syntax, creating diagrams, or visualizing complex concepts and workflows.
Google Docs/Sheets management via ACSet condensation. Transforms documents into GF(3)-typed Interactions, tracks comments/cells, detects saturation when all comments resolved. Use for document workflows, spreadsheet automation, or applying ANIMA principles to Workspace documents.
Generate bifurcation diagrams for dynamical systems. Use when visualizing parameter-dependent behavior transitions.
Use when interpreting Culture Index surveys, CI profiles, behavioral assessments, or personality data. Supports individual interpretation, team composition (gas/brake/glue), burnout detection, profile comparison, hiring profiles, manager coaching, interview transcript analysis for trait prediction, candidate debrief, onboarding planning, and conflict mediation. Handles PDF vision or JSON input.
Attributed C-Sets as algebraic databases. Category-theoretic data structures generalizing graphs and dataframes with Gay.jl color integration.
Framework for retrospectives at any time scale (daily, weekly, monthly, quarterly, yearly). Trigger with "daily retro", "weekly retro", "monthly retro", "[month] retro" (e.g., "december retro", "january retro"), "retro for [month]", or "end of month review". Answers three questions - what worked, what didn't, how to improve. Inputs vary by scale - daily uses raw logs, weekly uses daily summaries, monthly uses weekly retros, etc. Fractal compression pattern.
Generate comprehensive work summaries from daily notes or conversation context. Use for performance reviews, project retrospectives, or documenting completed work with timeline and impact.
Generate comprehensive work summaries from daily notes or conversation context. Use for performance reviews, project retrospectives, or documenting completed work with timeline and impact.
Framework for creating session summary documents. Use when user says "daily summary", "generate summary", or similar. Provides structured markdown template with date verification, filename patterns, and formatting guidelines. Supports range-based naming for sessions spanning calendar boundaries. Extend with personal metrics and domain-specific content.
Analyze, filter, transform, and convert log files using Kelora. Use for parsing logs, extracting patterns, investigating incidents, calculating metrics, or converting formats.
Identify and validate profitable business opportunities by analyzing market size (TAM/SAM/SOM), unit economics, competitive landscape, and PMF indicators. Generates comprehensive HTML reports with opportunity scorecards.
Generate an HTML dashboard from meeting analysis. Use when the user wants to create a visual summary, build a dashboard, generate a report, or visualize meeting insights.
Workflow for multi-step financial research requiring multiple data sources. Use for company comparisons, due diligence, comprehensive analysis, or complex financial questions.
Plot timestamped logs as graphs. Use when user wants to visualize log data, plot numeric values over time, count events, track time deltas between events, compare multiple log files, or get statistics from logs.
获取A股数据(baostock)并缓存到本地CSV文件,避免MCP返回大量数据占用上下文。触发场景:(1)获取超过100条的K线数据 (2)需要多次查询同一股票数据 (3)需要用grep/awk分析数据 (4)用户提到"保存数据"或"缓存数据
Analyze Mixpanel analytics data using the mixpanel_data Python library or mp CLI. Use when working with Mixpanel event data, user profiles, funnels, retention, cohorts, segmentation queries, JQL scripts, or SQL analysis on local DuckDB. Triggers on mentions of Mixpanel, event analytics, funnel analysis, retention curves, user behavior tracking, JQL queries, filter expressions, 'fetch data from Mixpanel', 'query Mixpanel with SQL', 'run DuckDB queries on events', 'analyze user behavior', 'export
Use when animating charts, graphs, dashboards, data transitions, or any information visualization work.
Structured reasoning tools for complex problem analysis with observation, hypothesis, analysis, and conclusion steps. Use when analyzing complex problems, debugging difficult issues, or making important decisions.
Hypothesis-Test Loops via REPL for Exploratory Abductive Inference with Gay.jl colors
Create D3.js charts and interactive data visualizations. Use when building bar charts, line charts, scatter plots, pie charts, force-directed graphs, geographic maps, or any custom data visualization.
Overview of protected quality thresholds and quick reference for all quality tools. Use when you need to understand quality metrics, run comprehensive quality checks, or learn which specialized skill to use. For specific issues, use dedicated skills (deptrac-fixer for Deptrac, complexity-management for PHPInsights, testing-workflow for coverage).
Create factual working journal entries in Notes/WorkingJournal/ after completing analysis work. Use when user asks to "summarize work", "document results", or "create working journal entry". Ensures code is committed, copies figures to attachments, and creates objective summaries with citations.
Generates concise summaries of log content using type-specific summarization templates
Analyzes and optimizes SQL/NoSQL queries for performance. Use when reviewing query performance, optimizing slow queries, analyzing EXPLAIN output, suggesting indexes, identifying N+1 problems, recommending query rewrites, or improving database access patterns. Supports PostgreSQL, MySQL, SQLite, MongoDB, Redis, DynamoDB, and Elasticsearch.
Create evidence synthesis matrices for systematic reviews. Use when: (1) Organizing extracted data, (2) Comparing study characteristics, (3) Identifying patterns across studies, (4) Preparing synthesis for manuscripts.
Guide selection and interpretation of statistical hypothesis tests. Use when: (1) Choosing appropriate test for research data, (2) Checking assumptions before analysis, (3) Interpreting test results correctly, (4) Reporting statistical findings, (5) Troubleshooting assumption violations.
Conduct sensitivity analyses to test robustness of findings. Use when: (1) Testing assumption violations, (2) Meta-analysis robustness, (3) Handling missing data, (4) Examining outliers.
Conduct quantitative synthesis through meta-analysis. Use when: (1) Combining effect sizes across studies, (2) Systematic review synthesis, (3) Calculating summary effects, (4) Assessing heterogeneity.
Conduct subgroup analyses to examine effect moderation. Use when: (1) Testing pre-specified moderators, (2) Exploring heterogeneity, (3) Identifying differential effects, (4) Meta-analysis synthesis.
Interpret statistical results correctly and comprehensively. Use when: (1) Writing results sections, (2) Discussing findings, (3) Avoiding common misinterpretations, (4) Reporting effect sizes and confidence intervals.
Calculate statistical power and required sample sizes for research studies. Use when: (1) Designing experiments to determine sample size, (2) Justifying sample size for grant proposals or protocols, (3) Evaluating adequacy of existing studies, (4) Meeting NIH rigor standards for pre-registration, (5) Conducting retrospective power analysis to interpret null results.
Generates data cleaning pipelines for pandas/polars with handling for missing values, duplicates, outliers, type conversions, and data validation. Use when user asks to "clean data", "generate data pipeline", "handle missing values", or "remove duplicates from dataset".
Analyzes and optimizes database queries for PostgreSQL, MySQL, MongoDB with EXPLAIN plans, index suggestions, and N+1 query detection. Use when user asks to "optimize query", "analyze EXPLAIN plan", "fix slow queries", or "suggest database indexes".
