Skip to content

DuckDB CLI skills

Verified

DuckDB CLI specialist for SQL analysis, data processing and file conversion. Use for SQL queries, CSV/Parquet/JSON analysis, database queries, or data conversion. Triggers on "duckdb", "sql", "query", "data analysis", "parquet", "convert data".

1,596

Install

Claude Code

Add to .claude/skills/

About This Skill

# DuckDB CLI Specialist

Helps with data analysis, SQL queries and file conversion via DuckDB CLI.

Quick Start

Read data files directly with SQL ```bash # CSV duckdb -c "SELECT * FROM 'data.csv' LIMIT 10"

# Parquet duckdb -c "SELECT * FROM 'data.parquet'"

# Multiple files with glob duckdb -c "SELECT * FROM read_parquet('logs/*.parquet')"

# JSON duckdb -c "SELECT * FROM read_json_auto('data.json')" ```

Open persistent databases ```bash # Create/open database duckdb my_database.duckdb

# Read-only mode duckdb -readonly existing.duckdb ```

Command Line Arguments

Output formats (as flags) | Flag | Format | |------|--------| | `-csv` | Comma-separated | | `-json` | JSON array | | `-table` | ASCII table | | `-markdown` | Markdown table | | `-html` | HTML table | | `-line` | One value per line |

Execution arguments | Argument | Description | |----------|-------------| | `-c COMMAND` | Run SQL and exit | | `-f FILENAME` | Run script from file | | `-init FILE` | Use alternative to ~/.duckdbrc | | `-readonly` | Open in read-only mode | | `-echo` | Show commands before execution | | `-bail` | Stop on first error | | `-header` / `-noheader` | Show/hide column headers | | `-nullvalue TEXT` | Text for NULL values | | `-separator SEP` | Column separator |

Data Conversion

CSV to Parquet ```bash duckdb -c "COPY (SELECT * FROM 'input.csv') TO 'output.parquet' (FORMAT PARQUET)" ```

Parquet to CSV ```bash duckdb -c "COPY (SELECT * FROM 'input.parquet') TO 'output.csv' (HEADER, DELIMITER ',')" ```

JSON to Parquet ```bash duckdb -c "COPY (SELECT * FROM read_json_auto('input.json')) TO 'output.parquet' (FORMAT PARQUET)" ```

Convert with filtering ```bash duckdb -c "COPY (SELECT * FROM 'data.csv' WHERE amount > 1000) TO 'filtered.parquet' (FORMAT PARQUET)" ```

Dot Commands

Schema inspection | Command | Description | |---------|-------------| | `.tables [pattern]` | Show tables (with LIKE pattern) | | `.schema [table]` | Show CREATE statements | | `.databases` | Show attached databases |

Output control | Command | Description | |---------|-------------| | `.mode FORMAT` | Change output format | | `.output file` | Send output to file | | `.once file` | Next output to file | | `.headers on/off` | Show/hide column headers | | `.separator COL ROW` | Set separators |

Queries | Command | Description | |---------|-------------| | `.timer on/off` | Show execution time | | `.echo on/off` | Show commands before execution | | `.bail on/off` | Stop on error | | `.read file.sql` | Run SQL from file |

Editing | Command | Description | |---------|-------------| | `.edit` or `\e` | Open query in external editor | | `.help [pattern]` | Show help |

Output Formats (18 available)

Data export - **csv** - Comma-separated for spreadsheets - **tabs** - Tab-separated - **json** - JSON array - **jsonlines** - Newline-delimited JSON (streaming)

Readable formats - **duckbox** (default) - Pretty ASCII with unicode box-drawing - **table** - Simple ASCII table - **markdown** - For documentation - **html** - HTML table - **latex** - For academic papers

Specialized - **insert TABLE** - SQL INSERT statements - **column** - Columns with adjustable width - **line** - One value per line - **list** - Pipe-separated - **trash** - Discard output

Keyboard Shortcuts (macOS/Linux)

Navigation | Shortcut | Action | |----------|--------| | `Home` / `End` | Start/end of line | | `Ctrl+Left/Right` | Jump word | | `Ctrl+A` / `Ctrl+E` | Start/end of buffer |

History | Shortcut | Action | |----------|--------| | `Ctrl+P` / `Ctrl+N` | Previous/next command | | `Ctrl+R` | Search history | | `Alt+<` / `Alt+>` | First/last in history |

Editing | Shortcut | Action | |----------|--------| | `Ctrl+W` | Delete word backward | | `Alt+D` | Delete word forward | | `Alt+U` / `Alt+L` | Uppercase/lowercase word | | `Ctrl+K` | Delete to end of line |

Autocomplete | Shortcut | Action | |----------|--------| | `Tab` | Autocomplete / next suggestion | | `Shift+Tab` | Previous suggestion | | `Esc+Esc` | Undo autocomplete |

Autocomplete

  • Context-aware autocomplete activated with `Tab`:
  • Keywords - SQL commands
  • Table names - Database objects
  • Column names - Fields and functions
  • File names - Path completion

Database Operations

Create table from file ```sql CREATE TABLE sales AS SELECT * FROM 'sales_2024.csv'; ```

Insert data ```sql INSERT INTO sales SELECT * FROM 'sales_2025.csv'; ```

Export table ```sql COPY sales TO 'backup.parquet' (FORMAT PARQUET); ```

Analysis Examples

Quick statistics ```sql SELECT COUNT(*) as count, AVG(amount) as average, SUM(amount) as total FROM 'transactions.csv'; ```

Grouping ```sql SELECT category, COUNT(*) as count, SUM(amount) as total FROM 'data.csv' GROUP BY category ORDER BY total DESC; ```

Join on files ```sql SELECT a.*, b.name FROM 'orders.csv' a JOIN 'customers.parquet' b ON a.customer_id = b.id; ```

Describe data ```sql DESCRIBE SELECT * FROM 'data.csv'; ```

Pipe and stdin

```bash # Read from stdin cat data.csv | duckdb -c "SELECT * FROM read_csv('/dev/stdin')"

# Pipe to another command duckdb -csv -c "SELECT * FROM 'data.parquet'" | head -20

# Write to stdout duckdb -c "COPY (SELECT * FROM 'data.csv') TO '/dev/stdout' (FORMAT CSV)" ```

Configuration

Save common settings in `~/.duckdbrc`: ```sql .timer on .mode duckbox .maxrows 50 .highlight on ```

Syntax highlighting colors ```sql .keyword green .constant yellow .comment brightblack .error red ```

External Editor

Open complex queries in your editor: ```sql .edit ```

Editor is chosen from: `DUCKDB_EDITOR` → `EDITOR` → `VISUAL` → `vi`

Safe Mode

  • Secure mode that restricts file access. When enabled:
  • No external file access
  • Disables `.read`, `.output`, `.import`, `.sh` etc.
  • Cannot be disabled in the same session

Tips

  • Use `LIMIT` on large files for quick preview
  • Parquet is faster than CSV for repeated queries
  • `read_csv_auto` and `read_json_auto` guess column types
  • Arguments are processed in order (like SQLite CLI)
  • WSL2 may show incorrect `memory_limit` values on some Ubuntu versions

Use Cases

  • Query local data files using DuckDB's high-performance SQL engine
  • Analyze Parquet, CSV, and JSON files directly with SQL without loading into a database
  • Run complex analytical queries on datasets that would be slow in SQLite or pandas
  • Join data from multiple file formats in a single SQL query
  • Export query results to Parquet or CSV for downstream processing

Pros & Cons

Pros

  • + DuckDB's columnar engine handles analytical queries orders of magnitude faster than SQLite
  • + Direct file querying eliminates ETL steps for Parquet, CSV, and JSON data
  • + Full SQL support including window functions, CTEs, and aggregations

Cons

  • - Requires DuckDB CLI to be installed on the system
  • - Only available on claude-code and openclaw platforms
  • - Not suitable for transactional workloads — designed for analytical queries only

Frequently Asked Questions

What does DuckDB CLI skills do?

DuckDB CLI specialist for SQL analysis, data processing and file conversion. Use for SQL queries, CSV/Parquet/JSON analysis, database queries, or data conversion. Triggers on "duckdb", "sql", "query", "data analysis", "parquet", "convert data".

What platforms support DuckDB CLI skills?

DuckDB CLI skills is available on Claude Code, OpenClaw.

What are the use cases for DuckDB CLI skills?

Query local data files using DuckDB's high-performance SQL engine. Analyze Parquet, CSV, and JSON files directly with SQL without loading into a database. Run complex analytical queries on datasets that would be slow in SQLite or pandas.

Stay Updated on Agent Skills

Get weekly curated skills + safety alerts