DataFrameWriter.
parquet
Saves the content of the DataFrame in Parquet format at the specified path.
DataFrame
New in version 1.4.0.
Changed in version 3.4.0: Supports Spark Connect.
the path in any Hadoop supported file system
specifies the behavior of the save operation when data already exists.
append: Append contents of this DataFrame to existing data.
append
overwrite: Overwrite existing data.
overwrite
ignore: Silently ignore this operation if data already exists.
ignore
error or errorifexists (default case): Throw an exception if data already exists.
error
errorifexists
names of partitioning columns
For the extra options, refer to Data Source Option for the version you use.
Examples
Write a DataFrame into a Parquet file and read it back.
>>> import tempfile >>> with tempfile.TemporaryDirectory() as d: ... # Write a DataFrame into a Parquet file ... spark.createDataFrame( ... [{"age": 100, "name": "Hyukjin Kwon"}] ... ).write.parquet(d, mode="overwrite") ... ... # Read the Parquet file as a DataFrame. ... spark.read.format("parquet").load(d).show() +---+------------+ |age| name| +---+------------+ |100|Hyukjin Kwon| +---+------------+