Databricks SQL Connector for Python
Project description
Databricks SQL Connector for Python
The Databricks SQL Connector for Python allows you to develop Python applications that connect to Databricks clusters and SQL warehouses. It is a Thrift-based client with no dependencies on ODBC or JDBC. It conforms to the Python DB API 2.0 specification.
This connector uses Arrow as the data-exchange format, and supports APIs (e.g. fetchmany_arrow
) to directly fetch Arrow tables. Arrow tables are wrapped in the ArrowQueue
class to provide a natural API to get several rows at a time. PyArrow is required to enable this and use these APIs, you can install it via pip install pyarrow
or pip install databricks-sql-connector[pyarrow]
.
You are welcome to file an issue here for general use cases. You can also contact Databricks Support here.
Requirements
Python 3.8 or above is required.
Documentation
For the latest documentation, see
Quickstart
Installing the core library
Install using pip install databricks-sql-connector
Installing the core library with PyArrow
Install using pip install databricks-sql-connector[pyarrow]
export DATABRICKS_HOST=********.databricks.com
export DATABRICKS_HTTP_PATH=/sql/1.0/endpoints/****************
Example usage:
import os
from databricks import sql
host = os.getenv("DATABRICKS_HOST")
http_path = os.getenv("DATABRICKS_HTTP_PATH")
connection = sql.connect(
server_hostname=host,
http_path=http_path)
cursor = connection.cursor()
cursor.execute('SELECT :param `p`, * FROM RANGE(10)', {"param": "foo"})
result = cursor.fetchall()
for row in result:
print(row)
cursor.close()
connection.close()
In the above example:
server-hostname
is the Databricks instance host name.http-path
is the HTTP Path either to a Databricks SQL endpoint (e.g. /sql/1.0/endpoints/1234567890abcdef), or to a Databricks Runtime interactive cluster (e.g. /sql/protocolv1/o/1234567890123456/1234-123456-slid123)
Note: This example uses Databricks OAuth U2M to authenticate the target Databricks user account and needs to open the browser for authentication. So it can only run on the user's machine.
SQLAlchemy
Starting from databricks-sql-connector
version 4.0.0 SQLAlchemy support has been extracted to a new library databricks-sqlalchemy
.
- Github repository databricks-sqlalchemy github
- PyPI databricks-sqlalchemy pypi
Quick SQLAlchemy guide
Users can now choose between using the SQLAlchemy v1 or SQLAlchemy v2 dialects with the connector core
- Install the latest SQLAlchemy v1 using
pip install databricks-sqlalchemy~=1.0
- Install SQLAlchemy v2 using
pip install databricks-sqlalchemy
Contributing
See CONTRIBUTING.md
License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file databricks_sql_connector-4.0.0b4.tar.gz
.
File metadata
- Download URL: databricks_sql_connector-4.0.0b4.tar.gz
- Upload date:
- Size: 315.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.9.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 465e50aed29d7947fcd664f09d8b056157bbe7988e1518cbec08a53921c48774 |
|
MD5 | fe34e382adc96e736cc890bdb86bbdb5 |
|
BLAKE2b-256 | f3058cd497124e73500ba7d36a13609002487714ff4397e4160ccf892962e98b |
File details
Details for the file databricks_sql_connector-4.0.0b4-py3-none-any.whl
.
File metadata
- Download URL: databricks_sql_connector-4.0.0b4-py3-none-any.whl
- Upload date:
- Size: 323.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.9.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e5f8e9bf3e73405190879b3c9d4962b2a19aaeeb5892ac3130189091e53d7104 |
|
MD5 | 2988479646573e64840f133f31fe5459 |
|
BLAKE2b-256 | 4eaadb32650427dea272601580f2df17fc20d054447c716cc4b5eef02f9a5a3a |