Showing 1-17 of 17 projects
Apache Airflow for workflow orchestration
Argo Workflows is a powerful open-source workflow engine for Kubernetes, enabling complex data processing and machine learning pipelines.
An open-source developer platform to power entire infra and turn scripts into webhooks, workflows, and UIs.
Apache DolphinScheduler is a modern data orchestration platform for creating high-performance workflows with low-code.
Curated list of resources about Apache Airflow, a popular workflow management platform.
A Docker-based Apache Airflow platform for building and managing data pipelines and workflows.
DataSphereStudio is a one-stop data application development and management portal covering data exchange, analysis, and visualization.
Elyra extends JupyterLab with an AI-centric approach for developing and deploying ML/AI pipelines.
A collection of Udacity data engineering projects showcasing various tools and technologies.
A set of Airflow DAGs to help maintain and manage the operation of an Airflow deployment.
This GitHub repository contains over 2,000 data engineering interview questions to help developers prepare.
An end-to-end data pipeline for building a data lake, data warehouse, and analytics platform from GoodReads data.
Declaratively construct Apache Airflow DAGs with YAML configuration files, simplifying complex data pipeline management.
An end-to-end data engineering project example showcasing tools and technologies for building data pipelines.
This repository provides best practices and examples for building ETL (Extract, Transform, Load) pipelines using Apache Airflow.
Run your dbt Core or dbt Fusion projects as Apache Airflow DAGs and Task Groups with a few lines of code.
A comprehensive knowledge hub for data engineering, machine learning, and MLOps tools and practices.
Get weekly updates on trending AI coding tools and projects.