A parquet file or dataset can be loaded using the read_parquet function. A parquet dataset is a directory with multiple parquet files, each of which is a partition belonging to the dataset.
Features
- The row range to iterate through, all rows by default. Applicable only when reading a single file
- Filter function to apply while loading only a subset of partitions from a dataset. The path to the partition is provided as a parameter
- Documentation available
- Examples available
- Lower Level Reader
- You can write any Tables.jl column-accessible table that contains columns of these types and their union with Missing: Int32, Int64, String, Bool, Float32, Float64
Categories
Data VisualizationLicense
MIT LicenseFollow Parquet.jl
You Might Also Like
Gen AI apps are built with MongoDB Atlas
MongoDB Atlas is the developer-friendly database used to build, scale, and run gen AI and LLM-powered apps—without needing a separate vector database. Atlas offers built-in vector search, global availability across 115+ regions, and flexible document modeling. Start building AI apps faster, all in one place.
Rate This Project
Login To Rate This Project
User Reviews
Be the first to post a review of Parquet.jl!