Spark Partition Discovery

Raymond Raymond
visibility 275 access_time 2 years ago language English

Spark supports partition discovery. All built in file sources (Text/CSV/JSON/ORC/Parquet) supports partition discovery and partition information inference. 

This data shows a example data set that is stored by two partition levels: month and country.

The following code snippet will read all the underlying parquet files:

df = spark.read.option("basePath","/data").parquet("/data")
info Last modified by Raymond 2 years ago copyright This page is subject to Site terms.
comment Comments
No comments yet.

Please log in or register to comment.

account_circle Log in person_add Register

Log in with external accounts

tag Tags
info Info
Image URL