GitHub - judeleonard/Prescriber-ETL-data-pipeline: An End-to-End ETL data pipeline that leverages pyspark parallel processing to process about 25 million rows of data coming from a SaaS application using Apache Airflow as an orchestration
![Streaming Data Pipeline to Transform, Store and Explore Healthcare Dataset With Apache Kafka API, Apache Spark, Apache Drill, JSON and MapR Database | HPE Developer Portal Streaming Data Pipeline to Transform, Store and Explore Healthcare Dataset With Apache Kafka API, Apache Spark, Apache Drill, JSON and MapR Database | HPE Developer Portal](https://hpe-developer-portal.s3.amazonaws.com/uploads/media/2020/12/example-streamline-processing-pipeline-1610604239860.png)
Streaming Data Pipeline to Transform, Store and Explore Healthcare Dataset With Apache Kafka API, Apache Spark, Apache Drill, JSON and MapR Database | HPE Developer Portal
![Apache Airflow(Open Source Tool for Data Pipelines) — Introduction and Installation | by Tushar Sharma | Medium Apache Airflow(Open Source Tool for Data Pipelines) — Introduction and Installation | by Tushar Sharma | Medium](https://miro.medium.com/v2/resize:fit:848/1*1lilEGyj8RKbCONTVGP7iw.png)
Apache Airflow(Open Source Tool for Data Pipelines) — Introduction and Installation | by Tushar Sharma | Medium
![How To Build A Real-time Data Pipeline For An Online Store Using Apache Beam, Pub/Sub, and SQL | by Aakash Rathor | Towards Data Science How To Build A Real-time Data Pipeline For An Online Store Using Apache Beam, Pub/Sub, and SQL | by Aakash Rathor | Towards Data Science](https://miro.medium.com/v2/resize:fit:1400/1*58TtOYNP71yhIoDcVDix4A.jpeg)