writers.arrow#
The Arrow Writer supports writing to Apache Arrow Feather and Parquet file types.
Dynamic Plugin
This stage requires a dynamic plugin to operate
Streamable Stage
This stage supports streaming operations
Example#
[
{
"type":"readers.las",
"filename":"inputfile.las"
},
{
"type":"writers.arrow",
"format":"feather",
"filename":"outputfile.feather"
}
]
[
{
"type":"readers.las",
"filename":"inputfile.las"
},
{
"type":"writers.arrow",
"format":"parquet",
"geoparquet":"true",
"filename":"outputfile.parquet"
}
]
Options#
- batch_size
Number of rows to write as a batch [Default: 65536*4 ]
- filename
Output file to write [Required]
- format
File type to write (feather, parquet) [Default: “feather”]
- geoarrow_dimension_name
Dimension name to write GeoArrow struct [Default: xyz]
- geoparquet
Write WKB column and GeoParquet metadata when writing parquet output
- write_pipeline_metadata
Write PDAL pipeline metadata into
PDAL:pipeline:metadata
ofgeoarrow_dimension_name
- where
An expression that limits points passed to a writer. Points that don’t pass the expression skip the stage but are available to subsequent stages in a pipeline. [Default: no filtering]
- where_merge
A strategy for merging points skipped by a ‘
where'
option when running in standard mode. Iftrue
, the skipped points are added to the first point view returned by the skipped filter. Iffalse
, skipped points are placed in their own point view. Ifauto
, skipped points are merged into the returned point view provided that only one point view is returned and it has the same point count as it did when the writer was run. [Default:auto
]