orc (DataFrameWriter)
Saves the content of the DataFrame in ORC format at the specified path.
Syntax
orc(path, mode=None, partitionBy=None, compression=None)
Parameters
Parameter | Type | Description |
|---|---|---|
| str | The path in any Hadoop-supported file system. |
| str, optional | The behavior when data already exists. Accepted values are |
| str or list, optional | Names of partitioning columns. |
| str, optional | The compression codec to use. |
Returns
None
Examples
Write a DataFrame into an ORC file and read it back.
Python
import tempfile
with tempfile.TemporaryDirectory(prefix="orc") as d:
spark.createDataFrame(
[{"age": 100, "name": "Alice"}]
).write.orc(d, mode="overwrite")
spark.read.format("orc").load(d).show()
# +---+------------+
# |age| name|
# +---+------------+
# |100|Alice|
# +---+------------+