polars.scan_ipc#
- polars.scan_ipc(
- source: str | Path | list[str] | list[Path],
- *,
- n_rows: int | None = None,
- cache: bool = True,
- rechunk: bool = True,
- row_count_name: str | None = None,
- row_count_offset: int = 0,
- storage_options: dict[str, Any] | None = None,
- memory_map: bool = True,
Lazily read from an Arrow IPC (Feather v2) file or multiple files via glob patterns.
This allows the query optimizer to push down predicates and projections to the scan level, thereby potentially reducing memory overhead.
- Parameters:
- source
Path to a IPC file.
- n_rows
Stop reading from IPC file after reading
n_rows
.- cache
Cache the result after reading.
- rechunk
Reallocate to contiguous memory when all chunks/ files are parsed.
- row_count_name
If not None, this will insert a row count column with give name into the DataFrame
- row_count_offset
Offset to start the row_count column (only use if the name is set)
- storage_options
Extra options that make sense for
fsspec.open()
or a particular storage connection. e.g. host, port, username, password, etc.- memory_map
Try to memory map the file. This can greatly improve performance on repeated queries as the OS may cache pages. Only uncompressed IPC files can be memory mapped.