Strategic Planning
by Shakes-tzd
Use HtmlGraph analytics to make smart work prioritization decisions. Activate when recommending work, finding bottlenecks, assessing risks, or analyzing project impact.
Skill Details
Repository Files
1 file in this skill directory
name: strategic-planning description: Use HtmlGraph analytics to make smart work prioritization decisions. Activate when recommending work, finding bottlenecks, assessing risks, or analyzing project impact.
Strategic Planning Skill
When to Activate This Skill
Trigger keywords:
- "what should I work on", "recommend", "prioritize"
- "bottleneck", "blocking", "stuck"
- "risk", "impact", "dependencies"
- "strategic", "roadmap", "plan"
Trigger situations:
- Starting a new session (what to work on?)
- Multiple tasks available (which is most important?)
- Progress seems slow (what's blocking us?)
- Planning major changes (what's the impact?)
Core Principle: Data-Driven Decisions
HtmlGraph provides analytics that consider:
- Dependencies - What blocks/enables other work
- Priority - Business importance
- Impact - How many tasks are unlocked
- Risk - Circular deps, complexity
- Parallelism - What can run concurrently
Quick Decision Framework
from htmlgraph import SDK
sdk = SDK(agent="claude")
# 1. What's blocking progress?
bottlenecks = sdk.find_bottlenecks(top_n=3)
if bottlenecks:
print("š§ BOTTLENECKS (fix these first):")
for bn in bottlenecks:
print(f" {bn['id']}: {bn['title']}")
print(f" Blocks {bn['blocks_count']} downstream tasks")
# 2. What should I work on?
recs = sdk.recommend_next_work(agent_count=1)
if recs:
best = recs[0]
print(f"\nš” RECOMMENDED: {best['title']}")
print(f" Score: {best['score']:.1f}")
print(f" Reasons: {', '.join(best['reasons'])}")
# 3. Can we parallelize?
parallel = sdk.get_parallel_work(max_agents=3)
print(f"\nā” Parallel capacity: {parallel['max_parallelism']} agents")
print(f" Ready now: {len(parallel['ready_now'])} tasks")
# 4. Any risks to watch?
risks = sdk.assess_risks()
if risks['high_risk_count'] > 0:
print(f"\nā ļø {risks['high_risk_count']} high-risk items")
Method Reference
sdk.find_bottlenecks(top_n=5)
Find tasks that block the most downstream work.
bottlenecks = sdk.find_bottlenecks(top_n=3)
# Returns list of:
{
"id": "feat-001",
"title": "Database Schema",
"status": "todo",
"priority": "high",
"blocks_count": 5, # How many tasks it blocks
"blocks": ["feat-002", "feat-003", ...] # Which tasks
}
Use when:
- Progress feels slow
- Many tasks are "blocked"
- Planning sprint priorities
sdk.recommend_next_work(agent_count=1)
Get scored recommendations considering all factors.
recs = sdk.recommend_next_work(agent_count=3)
# Returns list of:
{
"id": "feat-001",
"title": "Authentication",
"score": 85.5,
"reasons": [
"high_priority",
"unblocks_many",
"no_dependencies"
],
"priority": "high",
"status": "todo",
"blocks_count": 3
}
Scoring factors:
- Priority weight (critical=100, high=75, medium=50, low=25)
- Blocks count (Ć10 per blocked task)
- No dependencies bonus (+20)
- Bottleneck bonus (+30)
sdk.get_parallel_work(max_agents=5)
Find tasks that can run concurrently.
parallel = sdk.get_parallel_work(max_agents=5)
# Returns:
{
"max_parallelism": 4, # How many can run at once
"ready_now": ["f1", "f2", ...], # Level 0 (no deps)
"blocked": ["f3", "f4", ...], # Waiting on deps
"dependency_levels": [ # Topological layers
["f1", "f2"], # Level 0: no deps
["f3"], # Level 1: depends on Level 0
["f4", "f5"] # Level 2: depends on Level 1
]
}
Use when:
- Multiple agents available
- Want to speed up delivery
- Planning parallel sprints
sdk.assess_risks()
Check for project health issues.
risks = sdk.assess_risks()
# Returns:
{
"high_risk_count": 2,
"circular_dependencies": [], # Cycles in dep graph
"single_points_of_failure": [ # Tasks blocking many
{"id": "feat-001", "blocks": 5}
],
"stale_in_progress": [ # Stuck tasks
{"id": "feat-002", "days_stale": 7}
]
}
Use when:
- Before major releases
- Sprint planning
- Health checks
sdk.analyze_impact(feature_id)
Understand what completing a task unlocks.
impact = sdk.analyze_impact("feat-001")
# Returns:
{
"unlocks_count": 3,
"unlocks": ["feat-002", "feat-003", "feat-004"],
"transitive_impact": 7, # Total downstream tasks
"critical_path": True # On longest dependency chain
}
Use when:
- Deciding between tasks
- Explaining prioritization
- Finding high-leverage work
Decision Patterns
Pattern 1: Start of Session
sdk = SDK(agent="claude")
# Quick context
info = sdk.get_session_start_info()
print("š Project Status:")
print(f" In-progress: {info['status']['wip_count']}")
print(f" Bottlenecks: {len(info['bottlenecks'])}")
print(f" Parallel capacity: {info['parallel']['max_parallelism']}")
# What to work on
if info['recommendations']:
rec = info['recommendations'][0]
print(f"\nš” Start with: {rec['title']}")
Pattern 2: Something Is Blocked
# Find what's causing the block
bottlenecks = sdk.find_bottlenecks(top_n=5)
for bn in bottlenecks:
if bn['status'] == 'todo':
print(f"šÆ Unblock by completing: {bn['title']}")
print(f" This will enable {bn['blocks_count']} tasks")
break
Pattern 3: Planning Parallel Work
# Check if parallelization makes sense
parallel = sdk.get_parallel_work(max_agents=3)
risks = sdk.assess_risks()
if parallel['max_parallelism'] >= 2 and risks['high_risk_count'] == 0:
print("ā
Safe to parallelize")
print(f" Dispatch up to {parallel['max_parallelism']} agents")
# Get recommendations for each agent
recs = sdk.recommend_next_work(agent_count=parallel['max_parallelism'])
for i, rec in enumerate(recs):
print(f" Agent {i+1}: {rec['title']}")
else:
print("ā ļø Sequential execution recommended")
if risks['high_risk_count'] > 0:
print(f" Reason: {risks['high_risk_count']} high-risk items")
Pattern 4: Impact Analysis
# Compare two potential tasks
task_a = "feat-001"
task_b = "feat-002"
impact_a = sdk.analyze_impact(task_a)
impact_b = sdk.analyze_impact(task_b)
print(f"Task A unlocks: {impact_a['unlocks_count']} tasks")
print(f"Task B unlocks: {impact_b['unlocks_count']} tasks")
if impact_a['transitive_impact'] > impact_b['transitive_impact']:
print(f"š” Prioritize Task A (higher leverage)")
else:
print(f"š” Prioritize Task B (higher leverage)")
Integration with Smart Plan
The sdk.smart_plan() method combines these analytics:
plan = sdk.smart_plan(
description="Real-time collaboration",
create_spike=True,
timebox_hours=4
)
# Returns context with:
# - bottlenecks_count
# - high_risk_count
# - parallel_capacity
# - Created spike for research
Best Practices
DO
- Check bottlenecks first - High-leverage work
- Use recommendations - Considers all factors
- Assess risks before big changes - Avoid surprises
- Analyze impact - Understand consequences
- Check parallel capacity - Optimize throughput
DON'T
- Ignore blocked tasks - They signal bottlenecks
- Skip risk assessment - Before major releases
- Parallelize without analysis - May cause conflicts
- Work on low-impact tasks - When bottlenecks exist
Quick Reference
from htmlgraph import SDK
sdk = SDK(agent="claude")
# What's blocking us?
sdk.find_bottlenecks(top_n=5)
# What should I do?
sdk.recommend_next_work(agent_count=1)
# Can we parallelize?
sdk.get_parallel_work(max_agents=5)
# Any risks?
sdk.assess_risks()
# What does this unlock?
sdk.analyze_impact("feat-id")
# All-in-one session start
sdk.get_session_start_info()
# Smart planning
sdk.smart_plan("description")
Related Skills
Team Composition Analysis
This skill should be used when the user asks to "plan team structure", "determine hiring needs", "design org chart", "calculate compensation", "plan equity allocation", or requests organizational design and headcount planning for a startup.
Startup Financial Modeling
This skill should be used when the user asks to "create financial projections", "build a financial model", "forecast revenue", "calculate burn rate", "estimate runway", "model cash flow", or requests 3-5 year financial planning for a startup.
Startup Metrics Framework
This skill should be used when the user asks about "key startup metrics", "SaaS metrics", "CAC and LTV", "unit economics", "burn multiple", "rule of 40", "marketplace metrics", or requests guidance on tracking and optimizing business performance metrics.
Market Sizing Analysis
This skill should be used when the user asks to "calculate TAM", "determine SAM", "estimate SOM", "size the market", "calculate market opportunity", "what's the total addressable market", or requests market sizing analysis for a startup or business opportunity.
Anndata
This skill should be used when working with annotated data matrices in Python, particularly for single-cell genomics analysis, managing experimental measurements with metadata, or handling large-scale biological datasets. Use when tasks involve AnnData objects, h5ad files, single-cell RNA-seq data, or integration with scanpy/scverse tools.
Geopandas
Python library for working with geospatial vector data including shapefiles, GeoJSON, and GeoPackage files. Use when working with geographic data for spatial analysis, geometric operations, coordinate transformations, spatial joins, overlay operations, choropleth mapping, or any task involving reading/writing/analyzing vector geographic data. Supports PostGIS databases, interactive maps, and integration with matplotlib/folium/cartopy. Use for tasks like buffer analysis, spatial joins between dat
Market Research Reports
Generate comprehensive market research reports (50+ pages) in the style of top consulting firms (McKinsey, BCG, Gartner). Features professional LaTeX formatting, extensive visual generation with scientific-schematics and generate-image, deep integration with research-lookup for data gathering, and multi-framework strategic analysis including Porter's Five Forces, PESTLE, SWOT, TAM/SAM/SOM, and BCG Matrix.
Plotly
Interactive scientific and statistical data visualization library for Python. Use when creating charts, plots, or visualizations including scatter plots, line charts, bar charts, heatmaps, 3D plots, geographic maps, statistical distributions, financial charts, and dashboards. Supports both quick visualizations (Plotly Express) and fine-grained customization (graph objects). Outputs interactive HTML or static images (PNG, PDF, SVG).
Excel Analysis
Analyze Excel spreadsheets, create pivot tables, generate charts, and perform data analysis. Use when analyzing Excel files, spreadsheets, tabular data, or .xlsx files.
Neurokit2
Comprehensive biosignal processing toolkit for analyzing physiological data including ECG, EEG, EDA, RSP, PPG, EMG, and EOG signals. Use this skill when processing cardiovascular signals, brain activity, electrodermal responses, respiratory patterns, muscle activity, or eye movements. Applicable for heart rate variability analysis, event-related potentials, complexity measures, autonomic nervous system assessment, psychophysiology research, and multi-modal physiological signal integration.
