PySpark

https://spark.apache.org/docs/latest/api/python

Install

Install Ibis and dependencies for the PySpark backend:

Install with the pyspark extra:

pip install 'ibis-framework[pyspark]'

And connect:

import ibis

con = ibis.pyspark.connect()
1
Adjust connection parameters as needed.

Install for PySpark:

conda install -c conda-forge ibis-pyspark

And connect:

import ibis

con = ibis.pyspark.connect()
1
Adjust connection parameters as needed.

Install for PySpark:

mamba install -c conda-forge ibis-pyspark

And connect:

import ibis

con = ibis.pyspark.connect()
1
Adjust connection parameters as needed.

Connect

ibis.pyspark.connect

con = ibis.pyspark.connect(session=session)
Note

ibis.pyspark.connect is a thin wrapper around ibis.backends.pyspark.Backend.do_connect.

Note

The pyspark backend does not create SparkSession objects (unless you connect using a URL); you must create a SparkSession and pass that to ibis.pyspark.connect.

Connection Parameters

do_connect

do_connect(self, session=None, mode=None)

Create a PySpark Backend for use with Ibis.

Parameters
Name Type Description Default
session pyspark.sql.SparkSession | None A SparkSession instance None
mode ibis.backends.pyspark.ConnectionMode | None Can be either “batch” or “streaming”. If “batch”, every source, sink, and query executed within this connection will be interpreted as a batch workload. If “streaming”, every source, sink, and query executed within this connection will be interpreted as a streaming workload. None
Examples
>>> import ibis
>>> from pyspark.sql import SparkSession
>>> session = SparkSession.builder.getOrCreate()
>>> ibis.pyspark.connect(session)
<ibis.backends.pyspark.Backend at 0x...>

ibis.connect URL format

In addition to ibis.pyspark.connect, you can also connect to PySpark by passing a properly-formatted PySpark connection URL to ibis.connect:

con = ibis.connect(f"pyspark://{warehouse-dir}?spark.app.name=CountingSheep&spark.master=local[2]")

pyspark.Backend

add_operation

add_operation(self, operation)

Add a translation function to the backend for a specific operation.

Operations are defined in ibis.expr.operations, and a translation function receives the translator object and an expression as parameters, and returns a value depending on the backend.

compile

compile(self, expr, limit=None, params=None, pretty=False, **kwargs)

Compile an Ibis expression to a SQL string.

compute_stats

compute_stats(self, name, database=None, noscan=False)

Issue a COMPUTE STATISTICS command for a given table.

Parameters

Name Type Description Default
name str Table name required
database str | None Database name None
noscan bool If True, collect only basic statistics for the table (number of rows, size in bytes). False

connect

connect(self, *args, **kwargs)

Connect to the database.

Parameters

Name Type Description Default
*args Mandatory connection parameters, see the docstring of do_connect for details. ()
**kwargs Extra connection parameters, see the docstring of do_connect for details. {}

Notes

This creates a new backend instance with saved args and kwargs, then calls reconnect and finally returns the newly created and connected backend instance.

Returns

Type Description
ibis.backends.BaseBackend An instance of the backend

create_database

create_database(self, name, *, catalog=None, path=None, force=False)

Create a new Spark database.

Parameters

Name Type Description Default
name str Database name required
catalog str | None Catalog to create database in (defaults to current_catalog) None
path str | pathlib.Path | None Path where to store the database data; otherwise uses Spark default None
force bool Whether to append IF NOT EXISTS to the database creation SQL False

create_table

create_table(self, name, obj=None, *, schema=None, database=None, temp=None, overwrite=False, format='parquet')

Create a new table in Spark.

Parameters

Name Type Description Default
name str Name of the new table. required
obj ibis.expr.types.ibis.expr.types.Table | pandas.pandas.DataFrame | pyarrow.pyarrow.Table | polars.polars.DataFrame | polars.polars.LazyFrame | None If passed, creates table from SELECT statement results None
schema ibis.expr.schema.ibis.expr.schema.Schema | None Mutually exclusive with obj, creates an empty table with a schema None
database str | None Database name To specify a table in a separate catalog, you can pass in the catalog and database as a string "catalog.database", or as a tuple of strings ("catalog", "database"). None
temp bool | None Whether the new table is temporary (unsupported) None
overwrite bool If True, overwrite existing data False
format str Format of the table on disk 'parquet'

Returns

Type Description
Table The newly created table.

Examples

>>> con.create_table("new_table_name", table_expr)  # quartodoc: +SKIP

create_view

create_view(self, name, obj, *, database=None, overwrite=False)

Create a temporary Spark view from a table expression.

Parameters

Name Type Description Default
name str View name required
obj ibis.expr.types.ibis.expr.types.Table Expression to use for the view required
database str | None Database name None
overwrite bool Replace an existing view of the same name if it exists False

Returns

Type Description
Table The created view

disconnect

disconnect(self)

drop_database

drop_database(self, name, *, catalog=None, force=False)

Drop a Spark database.

Parameters

Name Type Description Default
name str Database name required
catalog str | None Catalog containing database to drop (defaults to current_catalog) None
force bool If False, Spark throws exception if database is not empty or database does not exist False

drop_table

drop_table(self, name, database=None, force=False)

drop_view

drop_view(self, name, *, database=None, schema=None, force=False)

execute

execute(self, expr, params=None, limit='default', **kwargs)

Execute an expression.

get_schema

get_schema(self, table_name, *, catalog=None, database=None)

Return a Schema object for the indicated table and database.

Parameters

Name Type Description Default
table_name str Table name. May be fully qualified required
catalog str | None Catalog to use None
database str | None Database to use to get the active database. None

Returns

Type Description
Schema An ibis schema

has_operation

has_operation(cls, operation)

insert

insert(self, table_name, obj, schema=None, database=None, overwrite=False)

Insert data into a table.

Ibis does not use the word schema to refer to database hierarchy.

A collection of table is referred to as a database. A collection of database is referred to as a catalog.

These terms are mapped onto the corresponding features in each backend (where available), regardless of whether the backend itself uses the same terminology.

Parameters

Name Type Description Default
table_name str The name of the table to which data needs will be inserted required
obj pandas.pandas.DataFrame | ibis.expr.types.ibis.expr.types.Table | list | dict The source data or expression to insert required
schema str | None [deprecated] The name of the schema that the table is located in None
database str | None Name of the attached database that the table is located in. For backends that support multi-level table hierarchies, you can pass in a dotted string path like "catalog.database" or a tuple of strings like ("catalog", "database"). None
overwrite bool If True then replace existing contents of table False

list_catalogs

list_catalogs(self, like=None)

List existing catalogs in the current connection.

Ibis does not use the word schema to refer to database hierarchy.

A collection of table is referred to as a database. A collection of database is referred to as a catalog.

These terms are mapped onto the corresponding features in each backend (where available), regardless of whether the backend itself uses the same terminology.

Parameters

Name Type Description Default
like str | None A pattern in Python’s regex format to filter returned database names. None

Returns

Type Description
list[str] The catalog names that exist in the current connection, that match the like pattern if provided.

list_databases

list_databases(self, like=None, catalog=None)

List existing databases in the current connection.

Ibis does not use the word schema to refer to database hierarchy.

A collection of table is referred to as a database. A collection of database is referred to as a catalog.

These terms are mapped onto the corresponding features in each backend (where available), regardless of whether the backend itself uses the same terminology.

Parameters

Name Type Description Default
like str | None A pattern in Python’s regex format to filter returned database names. None
catalog str | None The catalog to list databases from. If None, the current catalog is searched. None

Returns

Type Description
list[str] The database names that exist in the current connection, that match the like pattern if provided.

list_tables

list_tables(self, like=None, database=None)

List the tables in the database.

Parameters

Name Type Description Default
like str | None A pattern to use for listing tables. None
database str | None Database to list tables from. Default behavior is to show tables in the current catalog and database. To specify a table in a separate catalog, you can pass in the catalog and database as a string "catalog.database", or as a tuple of strings ("catalog", "database"). None

raw_sql

raw_sql(self, query, **kwargs)

read_csv

read_csv(self, source_list, table_name=None, **kwargs)

Register a CSV file as a table in the current database.

Parameters

Name Type Description Default
source_list str | list[str] | tuple[str] The data source(s). May be a path to a file or directory of CSV files, or an iterable of CSV files. required
table_name str | None An optional name to use for the created table. This defaults to a random generated name. None
kwargs typing.Any Additional keyword arguments passed to PySpark loading function. https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.DataFrameReader.csv.html {}

Returns

Type Description
ibis.expr.types.ibis.expr.types.Table The just-registered table

read_delta

read_delta(self, path, table_name=None, **kwargs)

Register a Delta Lake table as a table in the current database.

Parameters

Name Type Description Default
path str | pathlib.Path The path to the Delta Lake table. required
table_name str | None An optional name to use for the created table. This defaults to a random generated name. None
kwargs typing.Any Additional keyword arguments passed to PySpark. https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.DataFrameReader.load.html {}

Returns

Type Description
ibis.expr.types.ibis.expr.types.Table The just-registered table

read_json

read_json(self, source_list, table_name=None, **kwargs)

Register a JSON file as a table in the current database.

Parameters

Name Type Description Default
source_list str | collections.abc.Sequence[str] The data source(s). May be a path to a file or directory of JSON files, or an iterable of JSON files. required
table_name str | None An optional name to use for the created table. This defaults to a random generated name. None
kwargs typing.Any Additional keyword arguments passed to PySpark loading function. https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.DataFrameReader.json.html {}

Returns

Type Description
ibis.expr.types.ibis.expr.types.Table The just-registered table

read_parquet

read_parquet(self, path, table_name=None, **kwargs)

Register a parquet file as a table in the current database.

Parameters

Name Type Description Default
path str | pathlib.Path The data source. May be a path to a file or directory of parquet files. required
table_name str | None An optional name to use for the created table. This defaults to a random generated name. None
kwargs typing.Any Additional keyword arguments passed to PySpark. https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.DataFrameReader.parquet.html {}

Returns

Type Description
ibis.expr.types.ibis.expr.types.Table The just-registered table

reconnect

reconnect(self)

Reconnect to the database already configured with connect.

register

register(self, source, table_name=None, **kwargs)

Register a data source as a table in the current database.

Parameters

Name Type Description Default
source str | pathlib.Path | typing.Any The data source(s). May be a path to a file or directory of parquet/csv files, or an iterable of CSV files. required
table_name str | None An optional name to use for the created table. This defaults to a random generated name. None
**kwargs typing.Any Additional keyword arguments passed to PySpark loading functions for CSV or parquet. {}

Returns

Type Description
ibis.expr.types.ibis.expr.types.Table The just-registered table

register_options

register_options(cls)

Register custom backend options.

rename_table

rename_table(self, old_name, new_name)

Rename an existing table.

Parameters

Name Type Description Default
old_name str The old name of the table. required
new_name str The new name of the table. required

sql

sql(self, query, schema=None, dialect=None)

table

table(self, name, schema=None, database=None)

Construct a table expression.

Parameters

Name Type Description Default
name str Table name required
schema str | None [deprecated] Schema name None
database tuple[str, str] | str | None Database name None

Returns

Type Description
Table Table expression

to_csv

to_csv(self, expr, path, *, params=None, **kwargs)

Write the results of executing the given expression to a CSV file.

This method is eager and will execute the associated expression immediately.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Table The ibis expression to execute and persist to CSV. required
path str | pathlib.Path The data source. A string or Path to the CSV file. required
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Mapping of scalar parameter expressions to value. None
kwargs typing.Any Additional keyword arguments passed to pyarrow.csv.CSVWriter {}
https required

to_delta

to_delta(self, expr, path, **kwargs)

Write the results of executing the given expression to a Delta Lake table.

This method is eager and will execute the associated expression immediately.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Table The ibis expression to execute and persist to a Delta Lake table. required
path str | pathlib.Path The data source. A string or Path to the Delta Lake table. required
**kwargs typing.Any PySpark Delta Lake table write arguments. https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.DataFrameWriter.save.html {}

to_pandas

to_pandas(self, expr, *, params=None, limit=None, **kwargs)

Execute an Ibis expression and return a pandas DataFrame, Series, or scalar.

Note

This method is a wrapper around execute.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Expr Ibis expression to execute. required
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Mapping of scalar parameter expressions to value. None
limit int | str | None An integer to effect a specific row limit. A value of None means “no limit”. The default is in ibis/config.py. None
kwargs typing.Any Keyword arguments {}

to_pandas_batches

to_pandas_batches(self, expr, *, params=None, limit=None, chunk_size=1000000, **kwargs)

Execute an Ibis expression and return an iterator of pandas DataFrames.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Expr Ibis expression to execute. required
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Mapping of scalar parameter expressions to value. None
limit int | str | None An integer to effect a specific row limit. A value of None means “no limit”. The default is in ibis/config.py. None
chunk_size int Maximum number of rows in each returned DataFrame batch. This may have no effect depending on the backend. 1000000
kwargs typing.Any Keyword arguments {}

Returns

Type Description
collections.abc.Iterator[pandas.pandas.DataFrame] An iterator of pandas DataFrames.

to_parquet

to_parquet(self, expr, path, *, params=None, **kwargs)

Write the results of executing the given expression to a parquet file.

This method is eager and will execute the associated expression immediately.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Table The ibis expression to execute and persist to parquet. required
path str | pathlib.Path The data source. A string or Path to the parquet file. required
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Mapping of scalar parameter expressions to value. None
**kwargs typing.Any Additional keyword arguments passed to pyarrow.parquet.ParquetWriter {}
https required

to_polars

to_polars(self, expr, *, params=None, limit=None, **kwargs)

Execute expression and return results in as a polars DataFrame.

This method is eager and will execute the associated expression immediately.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Expr Ibis expression to export to polars. required
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Mapping of scalar parameter expressions to value. None
limit int | str | None An integer to effect a specific row limit. A value of None means “no limit”. The default is in ibis/config.py. None
kwargs typing.Any Keyword arguments {}

Returns

Type Description
dataframe A polars DataFrame holding the results of the executed expression.

to_pyarrow

to_pyarrow(self, expr, params=None, limit=None, **kwargs)

Execute expression and return results in as a pyarrow table.

This method is eager and will execute the associated expression immediately.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Expr Ibis expression to export to pyarrow required
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Mapping of scalar parameter expressions to value. None
limit int | str | None An integer to effect a specific row limit. A value of None means “no limit”. The default is in ibis/config.py. None
kwargs typing.Any Keyword arguments {}

Returns

Type Description
Table A pyarrow table holding the results of the executed expression.

to_pyarrow_batches

to_pyarrow_batches(self, expr, *, params=None, limit=None, chunk_size=1000000, **kwargs)

Execute expression and return an iterator of pyarrow record batches.

This method is eager and will execute the associated expression immediately.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Expr Ibis expression to export to pyarrow required
limit int | str | None An integer to effect a specific row limit. A value of None means “no limit”. The default is in ibis/config.py. None
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Mapping of scalar parameter expressions to value. None
chunk_size int Maximum number of rows in each returned record batch. 1000000

Returns

Type Description
RecordBatchReader Collection of pyarrow RecordBatchs.

to_torch

to_torch(self, expr, *, params=None, limit=None, **kwargs)

Execute an expression and return results as a dictionary of torch tensors.

Parameters

Name Type Description Default
expr ibis.expr.types.ibis.expr.types.Expr Ibis expression to execute. required
params collections.abc.Mapping[ibis.expr.types.ibis.expr.types.Scalar, typing.Any] | None Parameters to substitute into the expression. None
limit int | str | None An integer to effect a specific row limit. A value of None means no limit. None
kwargs typing.Any Keyword arguments passed into the backend’s to_torch implementation. {}

Returns

Type Description
dict[str, torch.torch.Tensor] A dictionary of torch tensors, keyed by column name.

truncate_table

truncate_table(self, name, database=None, schema=None)

Delete all rows from a table.

Ibis does not use the word schema to refer to database hierarchy.

A collection of tables is referred to as a database. A collection of database is referred to as a catalog. These terms are mapped onto the corresponding features in each backend (where available), regardless of whether the backend itself uses the same terminology.

Parameters

Name Type Description Default
name str Table name required
database str | None Name of the attached database that the table is located in. For backends that support multi-level table hierarchies, you can pass in a dotted string path like "catalog.database" or a tuple of strings like ("catalog", "database"). None
schema str | None [deprecated] Schema name None
Back to top