Anthropic Skills
2689 skills. Last updated 2026-03-15
Discover and use Anthropic Skills to extend Claude's capabilities with creative, technical, and enterprise workflows.
Expert patterns for Segment Customer Data Platform including Analytics.js, server-side tracking, tracking plans with Protocols, identity resolution, destinations configuration, and data governance best practices. Use when "segment, analytics.js, customer data platform, cdp, tracking plan, event tracking, identify track page, data routing, segment, cdp, analytics, tracking, data-pipeline, customer-data" mentioned.
Patterns for generating synthetic data for ML training, testing, and privacy. Covers LLM-based generation, tabular synthesis, and quality validation. Use when "synthetic data, generate training data, fake data generation, data augmentation, SDV, Gretel, test data, privacy-preserving data, " mentioned.
Comprehensive statistical analysis for research, experiments, and data science. Covers hypothesis testing, effect sizes, confidence intervals, Bayesian methods, regression, and advanced techniques. Emphasizes correct interpretation and avoiding common statistical mistakes. Use when ", " mentioned.
Data pipeline specialist for ETL design, data quality, CDC patterns, and batch/stream processingUse when "data pipeline, etl, cdc, data quality, batch processing, stream processing, data transformation, data warehouse, data lake, data validation, data-engineering, etl, cdc, batch, streaming, data-quality, dbt, airflow, dagster, data-pipeline, ml-memory" mentioned.
Causal inference specialist for causal discovery, counterfactual reasoning, and effect estimationUse when "causal inference, causal discovery, counterfactual, intervention effect, confounder, structural causal model, SCM, dowhy, causal graph, causal, dowhy, scm, dag, counterfactual, intervention, causalnex, confounding, ml-memory" mentioned.
Expert in measuring what matters in communities. Covers health metrics, engagement analytics, sentiment analysis, cohort tracking, and reporting. Knows that good data drives good decisions, and bad metrics drive bad behavior. Use when "community metrics, community analytics, measure community, community health, engagement metrics, community reporting, " mentioned.
Master the AI tools that transform financial operations. From bookkeeping to forecasting, manage money smarter and make data-driven financial decisions. Use when "finance, accounting, budgeting, forecasting, FP&A, bookkeeping, expense management, finance, accounting, budgeting, forecasting, fpa, expenses" mentioned.
Database design specialist for schema modeling, query optimization, indexing strategies, and data integrityUse when "database design, schema, indexes, query optimization, migrations, normalization, database scaling, foreign keys, data modeling, database, sql, postgres, mysql, mongodb, schema, indexes, migrations, normalization, optimization" mentioned.
Generate human-readable report from Intent files. Converts technical Intent specs into readable documents for stakeholders, team members, or documentation. Supports multiple output formats.
Generate and compare ydata-profiling EDA reports with sampling, consistent random seeds, and HTML outputs; often follows duckdb-parquet-lab-workflow when data is queried from Parquet.
Use DuckDB to query Parquet files, inspect metadata, join tables, and convert results to pandas for analysis; commonly precedes ydata-eda-profiling for EDA on extracted tables.
Comprehensive spreadsheet creation, editing, and analysis with support for formulas, formatting, data analysis, and visualization. Use when the assistant needs to work with spreadsheets (.xlsx, .xlsm, .csv, .tsv, etc) for: (1) Creating new spreadsheets with formulas and formatting, (2) Reading or analyzing data, (3) Modify existing spreadsheets while preserving formulas, (4) Data analysis and visualization in spreadsheets, or (5) Recalculating formulas
Advanced matrix theory expert covering spectral theory, matrix factorizations, and numerical linear algebra
Comprehensive statistics expert covering probability theory, distributions, hypothesis testing, regression, and Bayesian methods
Use when processing 50+ items, analyzing CSV/log files, executing code in sandbox, or searching for tools. Load for data processing tasks. Achieves 98%+ token savings via in-sandbox execution, progressive disclosure, and pagination. Supports heredocs for multi-line bash.
EDA, dashboards, Matplotlib, Seaborn, Plotly, and BI tools. Use for creating visualizations, exploratory analysis, or dashboards.
Automate data cleaning, transformation, and validation for ML tasks.
Reduce churn and improve retention through cohort analysis, at-risk user identification, win-back campaigns, and customer success strategies. Generate comprehensive HTML reports with retention curves, health scores, churn analysis, and 90-day implementation roadmaps.
Track and optimize application response times across API endpoints, database queries, and service calls. Use when monitoring performance or identifying bottlenecks. Trigger with phrases like "track response times", "monitor API performance", or "analyze latency".
Validate application performance against defined budgets to identify regressions early. Use when checking page load times, bundle sizes, or API response times against thresholds. Trigger with phrases like "validate performance budget", "check performance metrics", or "detect performance regression".
Create, select, and transform features to improve machine learning model
Comprehensive metrics dashboard strategy including North Star Metric definition, AARRR Pirate Metrics framework, product engagement tracking, 5 role-specific dashboards, alert configuration, data infrastructure planning, and 90-day implementation roadmap for data-driven decision making
Split datasets into training, validation, and testing sets for ML model
Provide comprehensive prioritized performance optimization recommendations for frontend, backend, and infrastructure. Use when analyzing bottlenecks or seeking improvement strategies. Trigger with phrases like "optimize performance", "improve speed", or "performance recommendations".
This skill applies Edward Tufte's data visualization principles from "The Visual Display of Quantitative Information" to create high-impact slides. Use when designing presentations, creating charts/graphs, reviewing slides for clarity, or transforming data into visual displays. Triggers on phrases like "make a slide", "create presentation", "design chart", "visualize data", "review my slides", or "make this more impactful".
Excel spreadsheet processing and analysis. Formats: .xlsx, .xlsm, .csv, .tsv. Capabilities: create spreadsheets, formulas (error-free), formatting, data analysis, charts, pivot tables, conditional formatting, data validation, template preservation. Actions: create, edit, analyze, visualize, recalculate spreadsheets. Keywords: Excel, spreadsheet, xlsx, csv, formula, VLOOKUP, SUMIF, pivot table, chart, graph, data analysis, formatting, conditional formatting, data validation, workbook, worksheet,
Assist users in creating quarterly connects that act as a strategic partner to guide employees through comprehensive quarterly reflections, helping craft insightful narratives for Quarterly Connection reviews that align with company values and career development goals.
UX research and design toolkit for Senior UX Designer/Researcher including data-driven persona generation, journey mapping, usability testing frameworks, and research synthesis. Use for user research, persona creation, journey mapping, and design validation.
Language-agnostic database best practices covering migrations, schema design, ORM patterns, query optimization, and testing strategies. Activate when working with database files, migrations, schema changes, SQL, ORM code, database tests, or when user mentions migrations, schema design, SQL optimization, NoSQL, database patterns, or connection pooling.
Generates professional performance analysis reports from SDL3 HammerEngine benchmark results including statistical analysis, comparison tables, visualizations, and recommendations. Use when preparing performance documentation, analyzing optimization efforts, or generating milestone/release reports.
