youtube image
From YouTube: Optimizing Delta Parquet Data Lakes for Apache Spark - Matthew Powers (Prognos)

Description

This talk will start by explaining the optimal file format, compression algorithm, and file size for plain vanilla Parquet data lakes. It discusses the small file problem and how you can compact the small files. Then we will talk about partitioning Parquet data lakes on disk and how to examine Spark physical plans when running queries on a partitioned lake. We will discuss why it's better to avoid PartitionFilters and directly grab partitions when querying partitioned lakes. We will explain why partitioned lakes tend to have a massive small file problem and why it's hard to compact a partitioned lake. Then we'll move on to Delta lakes and explain how they offer cool features on top of what's available in Parquet. We'll start with Delta 101 best practices and then move on to compacting with the OPTIMIZE command. We'll talk about creating partitioned Delta lake and how OPTIMIZE works on a partitioned lake. Then we'll talk about ZORDER indexes and how to incrementally update lakes with a ZORDER index. We'll finish with a discussion on adding a ZORDER index to a partitioned Delta data lake.

About: Databricks provides a unified data analytics platform, powered by Apache Spark™, that accelerates innovation by unifying data science, engineering and business.
Read more here: https://databricks.com/product/unified-data-analytics-platform

Connect with us:
Website: https://databricks.com
Facebook: https://www.facebook.com/databricksinc
Twitter: https://twitter.com/databricks
LinkedIn: https://www.linkedin.com/company/databricks
Instagram: https://www.instagram.com/databricksinc/ Databricks is proud to announce that Gartner has named us a Leader in both the 2021 Magic Quadrant for Cloud Database Management Systems and the 2021 Magic Quadrant for Data Science and Machine Learning Platforms. Download the reports here. https://databricks.com/databricks-named-leader-by-gartner