| name: datapipe |
| description: Data pipeline orchestrator for connecting sources, applying transformations, and routing to sinks on configurable schedules |
| binary: datapipe |
| auth: |
| type: env_var |
| key: DATAPIPE_API_KEY |
| commands: |
| - name: source list |
| description: List all configured data sources |
| args: |
| - name: type |
| type: enum |
| required: false |
| description: Filter by source type |
| values: ["postgres", "mysql", "s3", "api", "kafka", "mongodb"] |
| output_format: json |
| side_effects: false |
| example: "datapipe source list --type postgres" |
|
|
| - name: source connect |
| description: Register a new data source connection |
| args: |
| - name: name |
| type: string |
| required: true |
| description: Source name (unique identifier) |
| - name: type |
| type: enum |
| required: true |
| description: Source type |
| values: ["postgres", "mysql", "s3", "api", "kafka", "mongodb"] |
| - name: config |
| type: json |
| required: true |
| description: "Connection configuration as JSON (e.g. {\"host\": \"db.example.com\", \"port\": 5432, \"database\": \"analytics\"})" |
| output_format: json |
| side_effects: true |
| example: "datapipe source connect --name prod-db --type postgres --config '{\"host\": \"db.example.com\", \"port\": 5432, \"database\": \"analytics\"}'" |
|
|
| - name: transform create |
| description: Create a named transformation step with SQL logic |
| args: |
| - name: name |
| type: string |
| required: true |
| description: Transform name (unique identifier) |
| - name: sql |
| type: string |
| required: true |
| description: SQL transformation query |
| - name: source |
| type: string |
| required: true |
| description: Source name to read data from |
| - name: description |
| type: string |
| required: false |
| description: Human-readable description of the transformation |
| output_format: json |
| side_effects: true |
| example: "datapipe transform create --name daily-revenue --sql 'SELECT date, SUM(amount) as revenue FROM orders GROUP BY date' --source prod-db" |
|
|
| - name: transform run |
| description: Execute a transformation step |
| args: |
| - name: name |
| type: string |
| required: true |
| description: Transform name to execute |
| - name: dry-run |
| type: bool |
| required: false |
| description: Preview the output without writing results |
| - name: limit |
| type: int |
| required: false |
| description: Limit output rows (useful for preview) |
| output_format: json |
| side_effects: true |
| example: "datapipe transform run --name daily-revenue --dry-run --limit 10" |
|
|
| - name: sink create |
| description: Create a data sink destination |
| args: |
| - name: name |
| type: string |
| required: true |
| description: Sink name (unique identifier) |
| - name: type |
| type: enum |
| required: true |
| description: Sink type |
| values: ["s3", "bigquery", "warehouse", "postgres", "elasticsearch"] |
| - name: config |
| type: json |
| required: true |
| description: "Sink configuration as JSON (e.g. {\"bucket\": \"analytics-output\", \"prefix\": \"daily/\"})" |
| output_format: json |
| side_effects: true |
| example: "datapipe sink create --name analytics-lake --type s3 --config '{\"bucket\": \"analytics-output\", \"prefix\": \"daily/\"}'" |
|
|
| - name: pipeline create |
| description: Create a complete data pipeline connecting source, transforms, and sink with a schedule |
| args: |
| - name: name |
| type: string |
| required: true |
| description: Pipeline name (unique identifier) |
| - name: source |
| type: string |
| required: true |
| description: Source name |
| - name: transforms |
| type: string |
| required: true |
| description: Comma-separated list of transform names in execution order |
| - name: sink |
| type: string |
| required: true |
| description: Sink name |
| - name: schedule |
| type: string |
| required: true |
| description: Cron expression for scheduled execution |
| output_format: json |
| side_effects: true |
| example: "datapipe pipeline create --name daily-etl --source prod-db --transforms daily-revenue,add-dimensions --sink analytics-lake --schedule '0 2 * * *'" |
|
|
| - name: pipeline status |
| description: Check the status and execution history of a pipeline |
| args: |
| - name: name |
| type: string |
| required: true |
| description: Pipeline name |
| output_format: json |
| side_effects: false |
| example: "datapipe pipeline status --name daily-etl" |
|
|