Tdx Basic
by treasure-data
Executes tdx CLI commands for Treasure Data. Covers `tdx databases`, `tdx tables`, `tdx query`, `tdx auth setup`, context management with profiles/sessions, and output formats (JSON/TSV/table). Use when users need tdx command syntax, authentication setup, database/table exploration, or query execution.
Skill Details
Repository Files
1 file in this skill directory
name: tdx-basic
description: Executes tdx CLI commands for Treasure Data. Covers tdx databases, tdx tables, tdx query, tdx auth setup, context management with profiles/sessions, and output formats (JSON/TSV/table). Use when users need tdx command syntax, authentication setup, database/table exploration, or query execution.
tdx CLI - Basic Operations
Setup
npm install -g @treasuredata/tdx
# Interactive setup (recommended)
tdx auth setup
tdx auth setup --profile production # Multiple accounts
# Or manual: create ~/.config/tdx/.env
TDX_API_KEY=your-key-id/your-key-secret
# Verify
tdx auth
Context Management
Context priority: CLI flags > session > project tdx.json > profile > global config
# Session context
tdx use database mydb
tdx use site jp01
tdx profile use production # Switch profile
tdx status # View current context and auth
tdx use --clear # Clear session
Profile Management
tdx profile list # List all profiles
tdx profile create staging # Create new profile (interactive)
tdx profile set database=dev # Set profile config
tdx --profile staging query "..." # One-off with different profile
Profiles (~/.config/tdx/tdx.json)
{
"profiles": {
"production": { "site": "us01", "database": "analytics" },
"dev": { "site": "jp01", "database": "dev_db" }
}
}
Project Config (tdx.json in project root)
{
"site": "us01",
"database": "customer_analytics",
"parent_segment": "active_users"
}
Core Commands
Databases
tdx databases # List all
tdx databases "prod_*" # Filter with pattern
tdx databases --site jp01 # Specify site
tdx databases --json # JSON output
Sites: us01 (default), jp01, eu01, ap02
Tables
tdx use database mydb # Set context first
tdx tables # List tables
tdx tables "user_*" # Filter
tdx describe users # Schema
tdx show users --limit 10 # Preview data
# Pattern syntax
tdx tables "mydb.*" # All tables from mydb
tdx tables "*.users" # users table from all databases
Queries
tdx query "select * from mydb.users limit 10"
tdx query -f query.sql # From file
tdx query - # From stdin
echo "select 1" | tdx query -
tdx sg sql "Segment Name" | tdx query - # Pipe segment SQL
Output Formats
tdx databases --json # JSON
tdx query "..." --jsonl # JSON Lines (streaming)
tdx databases --tsv # TSV
tdx databases --output out.json # Save to file
Global Options
tdx <command> --help # Command help
--profile <name> # Use specific profile
--site <site> # us01, jp01, eu01, ap02
--json / --jsonl / --tsv # Output format
--output <file> # Save to file
--dry-run # Preview without executing
TD-Specific Conventions
- Table naming:
database_name.table_name - Time column: Unix timestamp (seconds since epoch), not datetime
- Time filtering: Use
td_interval(time, '-1d')ortd_time_range(time, 'start', 'end')for partition pruning - Timezone: UTC default; use
td_interval(time, '-1d', 'JST')for Japan
select time, from_unixtime(time) as datetime from mydb.events limit 1
Common Issues
| Issue | Solution |
|---|---|
| TDX_API_KEY not found | tdx auth setup or create ~/.config/tdx/.env |
| Database not found | Check site: tdx databases --site jp01 |
| Pattern not working | Quote patterns: tdx tables "prod_*" |
Project Folder Structure
tdx organizes resources into conventional folders:
my-project/
├── tdx.json # Project config (site, database, contexts)
├── parent_segments/ # Parent segment definitions
│ ├── customer-360.yml
│ └── demo-audience.yml
├── segments/ # Child segments and journeys (per parent)
│ └── customer-360/
│ ├── tdx.json # { "parent_segment": "Customer 360" }
│ ├── high-value.yml # Segment
│ ├── onboarding.yml # Journey (type: journey)
│ └── marketing/ # Folder organization
│ └── newsletter.yml
├── workflows/ # Digdag workflow projects
│ └── daily-etl/
│ ├── tdx.json # { "workflow_project": "daily-etl" }
│ └── main.dig
└── agents/ # LLM agents (per project)
└── my-llm-project/
├── tdx.json # { "llm_project": "My LLM Project" }
├── support-agent/
│ ├── agent.yml
│ └── prompt.md
└── knowledge_bases/
└── faq.yml
Each tdx.json stores context for its directory—commands run from any subdirectory use the nearest config.
Resources
- Documentation: https://tdx.treasuredata.com/
- Related: sql-skills/time-filtering, segment, journey, agent
Related Skills
Xlsx
Comprehensive spreadsheet creation, editing, and analysis with support for formulas, formatting, data analysis, and visualization. When Claude needs to work with spreadsheets (.xlsx, .xlsm, .csv, .tsv, etc) for: (1) Creating new spreadsheets with formulas and formatting, (2) Reading or analyzing data, (3) Modify existing spreadsheets while preserving formulas, (4) Data analysis and visualization in spreadsheets, or (5) Recalculating formulas
Clickhouse Io
ClickHouse database patterns, query optimization, analytics, and data engineering best practices for high-performance analytical workloads.
Clickhouse Io
ClickHouse database patterns, query optimization, analytics, and data engineering best practices for high-performance analytical workloads.
Analyzing Financial Statements
This skill calculates key financial ratios and metrics from financial statement data for investment analysis
Data Storytelling
Transform data into compelling narratives using visualization, context, and persuasive structure. Use when presenting analytics to stakeholders, creating data reports, or building executive presentations.
Kpi Dashboard Design
Design effective KPI dashboards with metrics selection, visualization best practices, and real-time monitoring patterns. Use when building business dashboards, selecting metrics, or designing data visualization layouts.
Dbt Transformation Patterns
Master dbt (data build tool) for analytics engineering with model organization, testing, documentation, and incremental strategies. Use when building data transformations, creating data models, or implementing analytics engineering best practices.
Sql Optimization Patterns
Master SQL query optimization, indexing strategies, and EXPLAIN analysis to dramatically improve database performance and eliminate slow queries. Use when debugging slow queries, designing database schemas, or optimizing application performance.
Clinical Decision Support
Generate professional clinical decision support (CDS) documents for pharmaceutical and clinical research settings, including patient cohort analyses (biomarker-stratified with outcomes) and treatment recommendation reports (evidence-based guidelines with decision algorithms). Supports GRADE evidence grading, statistical analysis (hazard ratios, survival curves, waterfall plots), biomarker integration, and regulatory compliance. Outputs publication-ready LaTeX/PDF format optimized for drug develo
Anndata
This skill should be used when working with annotated data matrices in Python, particularly for single-cell genomics analysis, managing experimental measurements with metadata, or handling large-scale biological datasets. Use when tasks involve AnnData objects, h5ad files, single-cell RNA-seq data, or integration with scanpy/scverse tools.
