Ingest and Write Columnar Data with Polars

Ingest and Write Columnar Data with Polars
.MP4, AVC, 1920x1080, 30 fps | English, AAC, 2 Ch | 1h 4m | 183 MB
Instructor: Surbhi Sharma
What you'll learn
Reliable data ingestion is one of the most critical and challenging aspects of building modern data pipelines. Raw files often arrive in different formats, schemas can drift, and poorly designed write patterns can break downstream analytics workflows. In this course, Ingest and Write Columnar Data with Polars, you'll gain the ability to design reliable and scalable data ingestion workflows using Polars.
First, you'll explore how to ingest common batch file formats such as CSV, JSON, and Parquet while defining explicit schemas and validation checks to prevent data quality issues. Next, you'll discover how to build scalable ingestion strategies for partitioned datasets, implement incremental file discovery, and normalize raw inputs into consistent column contracts for reliable processing. Finally, you'll learn how to write pipeline-friendly columnar outputs using formats such as Parquet, implement safe write patterns, and validate outputs to ensure downstream systems receive consistent datasets.
When you're finished with this course, you'll have the skills and knowledge of Polars-based data ingestion and writing techniques needed to build reliable, scalable, and analytics-ready data pipelines.
Homepage
