Duckdb En
by openclaw
DuckDB CLI specialist for SQL analysis, data processing and file conversion. Use for SQL queries, CSV/Parquet/JSON analysis, database queries, or data conversion. Triggers on "duckdb", "sql", "query", "data analysis", "parquet", "convert data".
Skill Details
Repository Files
3 files in this skill directory
name: duckdb-en description: DuckDB CLI specialist for SQL analysis, data processing and file conversion. Use for SQL queries, CSV/Parquet/JSON analysis, database queries, or data conversion. Triggers on "duckdb", "sql", "query", "data analysis", "parquet", "convert data".
DuckDB CLI Specialist
Helps with data analysis, SQL queries and file conversion via DuckDB CLI.
Quick Start
Read data files directly with SQL
# CSV
duckdb -c "SELECT * FROM 'data.csv' LIMIT 10"
# Parquet
duckdb -c "SELECT * FROM 'data.parquet'"
# Multiple files with glob
duckdb -c "SELECT * FROM read_parquet('logs/*.parquet')"
# JSON
duckdb -c "SELECT * FROM read_json_auto('data.json')"
Open persistent databases
# Create/open database
duckdb my_database.duckdb
# Read-only mode
duckdb -readonly existing.duckdb
Command Line Arguments
Output formats (as flags)
| Flag | Format |
|---|---|
-csv |
Comma-separated |
-json |
JSON array |
-table |
ASCII table |
-markdown |
Markdown table |
-html |
HTML table |
-line |
One value per line |
Execution arguments
| Argument | Description |
|---|---|
-c COMMAND |
Run SQL and exit |
-f FILENAME |
Run script from file |
-init FILE |
Use alternative to ~/.duckdbrc |
-readonly |
Open in read-only mode |
-echo |
Show commands before execution |
-bail |
Stop on first error |
-header / -noheader |
Show/hide column headers |
-nullvalue TEXT |
Text for NULL values |
-separator SEP |
Column separator |
Data Conversion
CSV to Parquet
duckdb -c "COPY (SELECT * FROM 'input.csv') TO 'output.parquet' (FORMAT PARQUET)"
Parquet to CSV
duckdb -c "COPY (SELECT * FROM 'input.parquet') TO 'output.csv' (HEADER, DELIMITER ',')"
JSON to Parquet
duckdb -c "COPY (SELECT * FROM read_json_auto('input.json')) TO 'output.parquet' (FORMAT PARQUET)"
Convert with filtering
duckdb -c "COPY (SELECT * FROM 'data.csv' WHERE amount > 1000) TO 'filtered.parquet' (FORMAT PARQUET)"
Dot Commands
Schema inspection
| Command | Description |
|---|---|
.tables [pattern] |
Show tables (with LIKE pattern) |
.schema [table] |
Show CREATE statements |
.databases |
Show attached databases |
Output control
| Command | Description |
|---|---|
.mode FORMAT |
Change output format |
.output file |
Send output to file |
.once file |
Next output to file |
.headers on/off |
Show/hide column headers |
.separator COL ROW |
Set separators |
Queries
| Command | Description |
|---|---|
.timer on/off |
Show execution time |
.echo on/off |
Show commands before execution |
.bail on/off |
Stop on error |
.read file.sql |
Run SQL from file |
Editing
| Command | Description |
|---|---|
.edit or \e |
Open query in external editor |
.help [pattern] |
Show help |
Output Formats (18 available)
Data export
- csv - Comma-separated for spreadsheets
- tabs - Tab-separated
- json - JSON array
- jsonlines - Newline-delimited JSON (streaming)
Readable formats
- duckbox (default) - Pretty ASCII with unicode box-drawing
- table - Simple ASCII table
- markdown - For documentation
- html - HTML table
- latex - For academic papers
Specialized
- insert TABLE - SQL INSERT statements
- column - Columns with adjustable width
- line - One value per line
- list - Pipe-separated
- trash - Discard output
Keyboard Shortcuts (macOS/Linux)
Navigation
| Shortcut | Action |
|---|---|
Home / End |
Start/end of line |
Ctrl+Left/Right |
Jump word |
Ctrl+A / Ctrl+E |
Start/end of buffer |
History
| Shortcut | Action |
|---|---|
Ctrl+P / Ctrl+N |
Previous/next command |
Ctrl+R |
Search history |
Alt+< / Alt+> |
First/last in history |
Editing
| Shortcut | Action |
|---|---|
Ctrl+W |
Delete word backward |
Alt+D |
Delete word forward |
Alt+U / Alt+L |
Uppercase/lowercase word |
Ctrl+K |
Delete to end of line |
Autocomplete
| Shortcut | Action |
|---|---|
Tab |
Autocomplete / next suggestion |
Shift+Tab |
Previous suggestion |
Esc+Esc |
Undo autocomplete |
Autocomplete
Context-aware autocomplete activated with Tab:
- Keywords - SQL commands
- Table names - Database objects
- Column names - Fields and functions
- File names - Path completion
Database Operations
Create table from file
CREATE TABLE sales AS SELECT * FROM 'sales_2024.csv';
Insert data
INSERT INTO sales SELECT * FROM 'sales_2025.csv';
Export table
COPY sales TO 'backup.parquet' (FORMAT PARQUET);
Analysis Examples
Quick statistics
SELECT
COUNT(*) as count,
AVG(amount) as average,
SUM(amount) as total
FROM 'transactions.csv';
Grouping
SELECT
category,
COUNT(*) as count,
SUM(amount) as total
FROM 'data.csv'
GROUP BY category
ORDER BY total DESC;
Join on files
SELECT a.*, b.name
FROM 'orders.csv' a
JOIN 'customers.parquet' b ON a.customer_id = b.id;
Describe data
DESCRIBE SELECT * FROM 'data.csv';
Pipe and stdin
# Read from stdin
cat data.csv | duckdb -c "SELECT * FROM read_csv('/dev/stdin')"
# Pipe to another command
duckdb -csv -c "SELECT * FROM 'data.parquet'" | head -20
# Write to stdout
duckdb -c "COPY (SELECT * FROM 'data.csv') TO '/dev/stdout' (FORMAT CSV)"
Configuration
Save common settings in ~/.duckdbrc:
.timer on
.mode duckbox
.maxrows 50
.highlight on
Syntax highlighting colors
.keyword green
.constant yellow
.comment brightblack
.error red
External Editor
Open complex queries in your editor:
.edit
Editor is chosen from: DUCKDB_EDITOR → EDITOR → VISUAL → vi
Safe Mode
Secure mode that restricts file access. When enabled:
- No external file access
- Disables
.read,.output,.import,.shetc. - Cannot be disabled in the same session
Tips
- Use
LIMITon large files for quick preview - Parquet is faster than CSV for repeated queries
read_csv_autoandread_json_autoguess column types- Arguments are processed in order (like SQLite CLI)
- WSL2 may show incorrect
memory_limitvalues on some Ubuntu versions
Related Skills
Xlsx
Comprehensive spreadsheet creation, editing, and analysis with support for formulas, formatting, data analysis, and visualization. When Claude needs to work with spreadsheets (.xlsx, .xlsm, .csv, .tsv, etc) for: (1) Creating new spreadsheets with formulas and formatting, (2) Reading or analyzing data, (3) Modify existing spreadsheets while preserving formulas, (4) Data analysis and visualization in spreadsheets, or (5) Recalculating formulas
Clickhouse Io
ClickHouse database patterns, query optimization, analytics, and data engineering best practices for high-performance analytical workloads.
Clickhouse Io
ClickHouse database patterns, query optimization, analytics, and data engineering best practices for high-performance analytical workloads.
Analyzing Financial Statements
This skill calculates key financial ratios and metrics from financial statement data for investment analysis
Data Storytelling
Transform data into compelling narratives using visualization, context, and persuasive structure. Use when presenting analytics to stakeholders, creating data reports, or building executive presentations.
Kpi Dashboard Design
Design effective KPI dashboards with metrics selection, visualization best practices, and real-time monitoring patterns. Use when building business dashboards, selecting metrics, or designing data visualization layouts.
Dbt Transformation Patterns
Master dbt (data build tool) for analytics engineering with model organization, testing, documentation, and incremental strategies. Use when building data transformations, creating data models, or implementing analytics engineering best practices.
Sql Optimization Patterns
Master SQL query optimization, indexing strategies, and EXPLAIN analysis to dramatically improve database performance and eliminate slow queries. Use when debugging slow queries, designing database schemas, or optimizing application performance.
Clinical Decision Support
Generate professional clinical decision support (CDS) documents for pharmaceutical and clinical research settings, including patient cohort analyses (biomarker-stratified with outcomes) and treatment recommendation reports (evidence-based guidelines with decision algorithms). Supports GRADE evidence grading, statistical analysis (hazard ratios, survival curves, waterfall plots), biomarker integration, and regulatory compliance. Outputs publication-ready LaTeX/PDF format optimized for drug develo
Anndata
This skill should be used when working with annotated data matrices in Python, particularly for single-cell genomics analysis, managing experimental measurements with metadata, or handling large-scale biological datasets. Use when tasks involve AnnData objects, h5ad files, single-cell RNA-seq data, or integration with scanpy/scverse tools.
