Spark Partition Discoveryimage

visibility 7 access_time 2mo languageEnglish

Spark supports partition discovery. All built in file sources (Text/CSV/JSON/ORC/Parquet) supports partition discovery and partition information inference. 

This data shows a example data set that is stored by two partition levels: month and country.

The following code snippet will read all the underlying parquet files:

df = spark.read.option("basePath","/data").parquet("/data")
info Last modified by Raymond 2mo copyright This page is subject to Site terms.

Please log in or register to comment.

account_circle Log in person_add Register

Log in with external accounts

comment Comments
No comments yet.
More from Kontext
Static and Dynamic Data Masking image
visibility 21
thumb_up 0
access_time 2mo
Static and Dynamic Data Masking
Go Programming [Banner] image
visibility 8
thumb_up 0
access_time 28d
Go Programming [Banner]