Project Feature Explainer

by GrishaAngelovGH

workflowtemplate

Expert guidance for explaining project features. Use this skill when you need to provide a comprehensive explanation of how a specific feature works, including summaries, deep dives, usage examples, and sequence/workflow diagrams. This skill has references directory which contains additional instructions `checklist.md`, `example-output.md` and `explanation-template.md` that MUST be used for every analysis.

Skill Details

Repository Files

4 files in this skill directory


name: project-feature-explainer description: Expert guidance for explaining project features. Use this skill when you need to provide a comprehensive explanation of how a specific feature works, including summaries, deep dives, usage examples, and sequence/workflow diagrams. This skill has references directory which contains additional instructions checklist.md, example-output.md and explanation-template.md that MUST be used for every analysis.

Project Feature Explainer Skill

This skill provides a systematic approach to analyzing and explaining a specific feature within a codebase.

Workflow

  1. Identify Entry Points: Locate the main functions, classes, or API endpoints that trigger the feature.
  2. Trace Dependencies: Identify the internal modules, services, or external APIs the feature relies on.
  3. Analyze Data Flow: Understand how data enters the feature, how it's transformed, and where it's stored or returned.
  4. Draft Explanation: Structure the explanation using the mandatory sections below.
  5. Verify: Cross-reference the draft with the references/checklist.md to ensure completeness.

Mandatory Output Structure

1. Feature Summary

A high-level overview (1-2 paragraphs) explaining what the feature does and why it exists.

2. Deep Dive (Technical Details)

A detailed breakdown of the internal implementation.

  • Key Components: List the main files/classes/functions involved.
  • Logic Flow: Step-by-step description of the execution path.
  • State Changes: Describe any side effects (e.g., database updates, cache invalidation).

3. Usage & Examples

Code snippets or CLI commands showing how to use the feature.

  • Include common parameters and expected outputs.
  • Provide a "Happy Path" example.

Guidelines

  • Be Concise: Focus on the "how" and "why" without repeating obvious code logic.
  • Reference Code: Use specific file paths and symbol names.
  • Contextualize: Explain how this feature fits into the broader system architecture.

Related Skills

Dask

Parallel/distributed computing. Scale pandas/NumPy beyond memory, parallel DataFrames/Arrays, multi-file processing, task graphs, for larger-than-RAM datasets and parallel workflows.

workflowdata

Scikit Survival

Comprehensive toolkit for survival analysis and time-to-event modeling in Python using scikit-survival. Use this skill when working with censored survival data, performing time-to-event analysis, fitting Cox models, Random Survival Forests, Gradient Boosting models, or Survival SVMs, evaluating survival predictions with concordance index or Brier score, handling competing risks, or implementing any survival analysis workflow with the scikit-survival library.

workflowtooldata

Polars

Fast DataFrame library (Apache Arrow). Select, filter, group_by, joins, lazy evaluation, CSV/Parquet I/O, expression API, for high-performance data analysis workflows.

workflowapidata

Scikit Survival

Comprehensive toolkit for survival analysis and time-to-event modeling in Python using scikit-survival. Use this skill when working with censored survival data, performing time-to-event analysis, fitting Cox models, Random Survival Forests, Gradient Boosting models, or Survival SVMs, evaluating survival predictions with concordance index or Brier score, handling competing risks, or implementing any survival analysis workflow with the scikit-survival library.

workflowtooldata

Dask

Distributed computing for larger-than-RAM pandas/NumPy workflows. Use when you need to scale existing pandas/NumPy code beyond memory or across clusters. Best for parallel file processing, distributed ML, integration with existing pandas code. For out-of-core analytics on single machine use vaex; for in-memory speed use polars.

codeworkflow

Anndata

Data structure for annotated matrices in single-cell analysis. Use when working with .h5ad files or integrating with the scverse ecosystem. This is the data format skill—for analysis workflows use scanpy; for probabilistic models use scvi-tools; for population-scale queries use cellxgene-census.

workflowtooldata

Matplotlib

Low-level plotting library for full customization. Use when you need fine-grained control over every plot element, creating novel plot types, or integrating with specific scientific workflows. Export to PNG/PDF/SVG for publication. For quick statistical plots use seaborn; for interactive plots use plotly; for publication-ready multi-panel figures with journal styling, use scientific-visualization.

workflow

Infographic Syntax Creator

Generate AntV Infographic syntax outputs. Use when asked to turn user content into the Infographic DSL (template selection, data structuring, theme), or to output `infographic <template>` plain syntax.

templatedata

Infographic Syntax Creator

Generate AntV Infographic syntax outputs. Use when asked to turn user content into the Infographic DSL (template selection, data structuring, theme), or to output `infographic <template>` plain syntax.

templatedata

Dashboard Design

USE THIS SKILL FIRST when user wants to create and design a dashboard, ESPECIALLY Vizro dashboards. This skill enforces a 3-step workflow (requirements, layout, visualization) that must be followed before implementation. For implementation and testing, use the dashboard-build skill after completing Steps 1-3.

designtestingworkflow

Skill Information

Category:Enterprise
Last Updated:1/8/2026