ETL and Data Pipelines with Shell, Airflow and Kafka

$49
ENROLL NOWCourse Overview
What You'll Learn
- In this course, you will learn about the different tools and techniques that are used with ETL and Data pipelines.
- By the end of this course, you will also know how to use Apache Airflow to build data pipelines as well be knowledgeable about the advantages of using this approach.
- You will also learn how to use Apache Kafka to build streaming pipelines as well as the core components of Kafka which include: brokers, topics, partitions, replications, producers, and consumers.
Delve into the two different approaches to converting raw data into analytics-ready data. One approach is the Extract, Transform, Load (ETL) process. The other contrasting approach is the Extract, Load, and Transform (ELT) process. ETL processes apply to data warehouses and data marts. ELT processes apply to data lakes, where the data is transformed on demand by the requesting/calling application. In this course, you will learn about the different tools and techniques that are used with ETL and Data pipelines. Both ETL and ELT extract data from source systems, move the data through the data pipeline, and store the data in destination systems. During this course, you will experience how ELT and ETL processing differ and identify use cases for both. You will identify methods and tools used for extracting the data, merging extracted data either logically or physically, and for loading data into data repositories. You will also define transformations to apply to source data to make the data credible, contextual, and accessible to data users. You will be able to outline some of the multiple methods for loading data into the destination system, verifying data quality, monitoring load failures, and the use of recovery mechanisms in case of failure. By the end of this course, you will also know how to use Apache Airflow to build data pipelines as well be knowledgeable about the advantages of using this approach. You will also learn how to use Apache Kafka to build streaming pipelines as well as the core components of Kafka which include: brokers, topics, partitions, replications, producers, and consumers. Finally, you will complete a shareable final project that enables you to demonstrate the skills you acquired in each module.
Course FAQs
Is this an accredited online course?
Accreditation for 'ETL and Data Pipelines with Shell, Airflow and Kafka' is determined by the provider, IBM. For online college courses or degree programs, we strongly recommend you verify the accreditation status directly on the provider's website to ensure it meets your requirements.
Can this course be used for continuing education credits?
Many of the courses listed on our platform are suitable for professional continuing education. However, acceptance for credit varies by state and licensing board. Please confirm with your board and {course.provider} that this specific course qualifies.
How do I enroll in this online school program?
To enroll, click the 'ENROLL NOW' button on this page. You will be taken to the official page for 'ETL and Data Pipelines with Shell, Airflow and Kafka' on the IBM online class platform, where you can complete your registration.




