🧪 Skills
Data Pipeline Toolkit
Create, schedule, and monitor ETL pipelines extracting from APIs, databases, files, and streams, with transform and load support to warehouses and APIs.
v1.0.0
Description
Data Pipeline Toolkit
Build ETL pipelines in minutes.
Features
- Extract - APIs, databases, files, streams
- Transform - Clean, filter, aggregate, join
- Load - Data warehouses, databases, APIs
- Scheduling - Cron-based or event-driven
- Monitoring - Alerts on failures
Quick Start
# Create pipeline
./pipeline.sh create my-pipeline
# Add extract step
./pipeline.sh extract my-pipeline api --url https://api.example.com
# Add transform
./pipeline.sh transform my-pipeline filter "age > 18"
# Add load
./pipeline.sh load my-pipeline postgres --connection $DB_URL
# Run
./pipeline.sh run my-pipeline
Sources
- REST APIs
- GraphQL
- SQL Databases
- CSV/JSON/Parquet files
- S3/Google Cloud Storage
- Kafka/SQS
Destinations
- PostgreSQL/MySQL
- Snowflake/BigQuery
- S3
- APIs
- Data warehouses
Requirements
- Python 3.8+
- Docker (optional)
Author
Sunshine-del-ux
Reviews (0)
Sign in to write a review.
No reviews yet. Be the first to review!
Comments (0)
No comments yet. Be the first to share your thoughts!