Data Quality Audit
by nimrodfisher
Comprehensive data quality assessment against defined business rules and constraints. Use when validating data against expected schemas, checking referential integrity across tables, or auditing data pipeline outputs before production use.
Skill Details
Repository Files
1 file in this skill directory
name: data-quality-audit description: Comprehensive data quality assessment against defined business rules and constraints. Use when validating data against expected schemas, checking referential integrity across tables, or auditing data pipeline outputs before production use.
Data Quality Audit
Quick Start
This skill helps you comprehensive data quality assessment against defined business rules and constraints.
Context Requirements
Before proceeding, I need:
- Schema relationships: Key information needed for this analysis
- Business rules by entity: Key information needed for this analysis
- Acceptable error rates: Key information needed for this analysis
- Critical vs non-critical fields: Key information needed for this analysis
Context Gathering
If any required context is missing from our conversation, I'll ask for it using these prompts:
For Schema relationships:
"To proceed with data quality audit, I need to understand schema relationships.
Please provide:
- [Specific detail 1 about schema relationships]
- [Specific detail 2 about schema relationships]
- [Optional context that would help]"
For Business rules by entity:
"To proceed with data quality audit, I need to understand business rules by entity.
Please provide:
- [Specific detail 1 about business rules by entity]
- [Specific detail 2 about business rules by entity]
- [Optional context that would help]"
For Acceptable error rates:
"To proceed with data quality audit, I need to understand acceptable error rates.
Please provide:
- [Specific detail 1 about acceptable error rates]
- [Specific detail 2 about acceptable error rates]
- [Optional context that would help]"
Handling Partial Context
If you can only provide some of the context:
- I'll proceed with what's available and note limitations
- I'll use industry standard defaults where appropriate
- I'll ask clarifying questions as needed during the analysis
Workflow
Step 1: Validate Context
Before starting, I'll confirm:
- All required context is available or has reasonable defaults
- The scope and objectives are clear
- Expected outputs align with your needs
Step 2: Execute Core Analysis
Following best practices for data quality audit, I'll:
- Initial assessment - Review provided context and data
- Systematic execution - Follow structured methodology
- Quality checks - Validate intermediate results
- Progressive disclosure - Share findings at logical checkpoints
Step 3: Synthesize Findings
I'll present results in a clear, actionable format:
- Key findings prioritized by importance
- Supporting evidence and visualizations
- Recommendations with implementation guidance
- Limitations and assumptions documented
Step 4: Iterate Based on Feedback
After presenting initial findings:
- Address questions and dive deeper where needed
- Refine analysis based on your feedback
- Provide additional context or alternative approaches
Context Validation
Before executing the full workflow, I verify:
- Context is sufficient for meaningful analysis
- No contradictions in provided information
- Scope is well-defined and achievable
- Expected outputs are clear
Output Template
Data Quality Audit Analysis
Generated: [timestamp]
## Context Summary
- [Key context item 1]
- [Key context item 2]
- [Key context item 3]
## Methodology
[Brief description of approach taken]
## Key Findings
1. **Finding 1**: [Observation] - [Implication]
2. **Finding 2**: [Observation] - [Implication]
3. **Finding 3**: [Observation] - [Implication]
## Detailed Analysis
[In-depth analysis with supporting evidence]
## Recommendations
1. **Recommendation 1**: [Action] - [Expected outcome]
2. **Recommendation 2**: [Action] - [Expected outcome]
## Limitations & Assumptions
- [Limitation or assumption 1]
- [Limitation or assumption 2]
## Next Steps
1. [Suggested follow-up action 1]
2. [Suggested follow-up action 2]
Common Context Gaps & Solutions
Scenario: User requests data quality audit without providing context → Response: "I can help with data quality audit! To provide the most relevant analysis, I need [key context items]. Can you share [specific ask]?"
Scenario: Partial context provided → Response: "I have [available context]. I'll proceed with [what's possible] and will note where additional context would improve the analysis."
Scenario: Unclear objectives
→ Response: "To ensure my analysis meets your needs, can you clarify: What decisions will this inform? What format would be most useful?"
Scenario: Domain-specific terminology → Response: "I want to make sure I understand your terminology correctly. When you say [term], do you mean [interpretation]?"
Advanced Options
Once basic analysis is complete, I can offer:
- Deeper investigation - Drill into specific findings
- Alternative approaches - Different analytical lenses
- Sensitivity analysis - Test key assumptions
- Comparative analysis - Benchmark against alternatives
- Visualization options - Different ways to present findings
Just ask if you'd like to explore any of these directions!
Integration with Other Skills
This skill works well in combination with:
- [Related skill 1] - for [complementary analysis]
- [Related skill 2] - for [next step in workflow]
- [Related skill 3] - for [alternative perspective]
Let me know if you'd like to chain multiple analyses together.
Related Skills
Xlsx
Comprehensive spreadsheet creation, editing, and analysis with support for formulas, formatting, data analysis, and visualization. When Claude needs to work with spreadsheets (.xlsx, .xlsm, .csv, .tsv, etc) for: (1) Creating new spreadsheets with formulas and formatting, (2) Reading or analyzing data, (3) Modify existing spreadsheets while preserving formulas, (4) Data analysis and visualization in spreadsheets, or (5) Recalculating formulas
Clickhouse Io
ClickHouse database patterns, query optimization, analytics, and data engineering best practices for high-performance analytical workloads.
Clickhouse Io
ClickHouse database patterns, query optimization, analytics, and data engineering best practices for high-performance analytical workloads.
Analyzing Financial Statements
This skill calculates key financial ratios and metrics from financial statement data for investment analysis
Data Storytelling
Transform data into compelling narratives using visualization, context, and persuasive structure. Use when presenting analytics to stakeholders, creating data reports, or building executive presentations.
Kpi Dashboard Design
Design effective KPI dashboards with metrics selection, visualization best practices, and real-time monitoring patterns. Use when building business dashboards, selecting metrics, or designing data visualization layouts.
Dbt Transformation Patterns
Master dbt (data build tool) for analytics engineering with model organization, testing, documentation, and incremental strategies. Use when building data transformations, creating data models, or implementing analytics engineering best practices.
Sql Optimization Patterns
Master SQL query optimization, indexing strategies, and EXPLAIN analysis to dramatically improve database performance and eliminate slow queries. Use when debugging slow queries, designing database schemas, or optimizing application performance.
Anndata
This skill should be used when working with annotated data matrices in Python, particularly for single-cell genomics analysis, managing experimental measurements with metadata, or handling large-scale biological datasets. Use when tasks involve AnnData objects, h5ad files, single-cell RNA-seq data, or integration with scanpy/scverse tools.
Xlsx
Spreadsheet toolkit (.xlsx/.csv). Create/edit with formulas/formatting, analyze data, visualization, recalculate formulas, for spreadsheet processing and analysis.
