Skip to main content

An easily customizable SQL parser and transpiler

Project description

SQLGlot

SQLGlot is a no dependency Python SQL parser and transpiler. It can be used to format SQL or translate between different dialects like Presto, Spark, and Hive. It aims to read a wide variety of SQL inputs and output syntatically correct SQL in the targeted dialects.

It is up to 3x faster than sqlparse.

You can easily customize the parser to support UDF's across dialects as well through the transform API.

Syntax errors are highlighted and dialect incompatibilities can warn or raise depending on configurations.

Install

From PyPI

pip3 install sqlglot

Or with a local checkout

pip3 install -e .

Examples

Easily translate from one dialect to another. For example, date/time functions vary from dialects and can be hard to deal with.

import sqlglot
sqlglot.transpile("SELECT EPOCH_MS(1618088028295)", read='duckdb', write='hive')
SELECT TO_UTC_TIMESTAMP(FROM_UNIXTIME(1618088028295 / 1000, 'yyyy-MM-dd HH:mm:ss'), 'UTC')

Formatting and Transpiling

Read in a SQL statement with a CTE and CASTING to a REAL and then transpiling to Spark.

Spark uses backticks as identifiers and the REAL type is transpiled to FLOAT.

import sqlglot

sql = """WITH baz AS (SELECT a, c FROM foo WHERE a = 1) SELECT f.a, b.b, baz.c, CAST("b"."a" AS REAL) d FROM foo f JOIN bar b ON f.a = b.a LEFT JOIN baz ON f.a = baz.a"""
sqlglot.transpile(sql, write='spark', identify=True, pretty=True)[0])
WITH baz AS (
    SELECT
      `a`,
      `c`
    FROM `foo`
    WHERE
      `a` = 1
)
SELECT
  `f`.`a`,
  `b`.`b`,
  `baz`.`c`,
  CAST(`b`.`a` AS FLOAT) AS d
FROM `foo` AS f
JOIN `bar` AS b ON
  `f`.`a` = `b`.`a`
LEFT JOIN `baz` ON
  `f`.`a` = `baz`.`a`

Custom Transforms

A simple transform on types can be accomplished by providing a dict of Expression/TokenType => lambda/string

from sqlglot import *

transpile("SELECT CAST(a AS INT) FROM x", transforms={TokenType.INT: 'SPECIAL INT'})[0]
SELECT CAST(a AS SPECIAL INT) FROM x

More complicated transforms can be accomplished by using the Tokenizer, Parser, and Generator directly.

In this example, we want to parse a UDF SPECIAL_UDF and then output another version called SPECIAL_UDF_INVERSE with the arguments switched.

from sqlglot import *
from sqlglot.expressions import Func

class SpecialUDF(Func):
    arg_types = {'a': True, 'b': True}

tokens = Tokenizer().tokenize("SELECT SPECIAL_UDF(a, b) FROM x")

Here is the output of the tokenizer.

[
    <Token token_type: TokenType.SELECT, text: SELECT, line: 0, col: 0>,
    <Token token_type: TokenType.VAR, text: SPECIAL_UDF, line: 0, col: 7>,
    <Token token_type: TokenType.L_PAREN, text: (, line: 0, col: 18>,
    <Token token_type: TokenType.VAR, text: a, line: 0, col: 19>,
    <Token token_type: TokenType.COMMA, text: ,, line: 0, col: 20>,
    <Token token_type: TokenType.VAR, text: b, line: 0, col: 22>,
    <Token token_type: TokenType.R_PAREN, text: ), line: 0, col: 23>,
    <Token token_type: TokenType.FROM, text: FROM, line: 0, col: 25>,
    <Token token_type: TokenType.VAR, text: x, line: 0, col: 30>,
]

expression = Parser(functions={
    'SPECIAL_UDF': lambda args: SpecialUDF(a=args[0], b=args[1]),
}).parse(tokens)[0]

The expression tree produced by the parser.

(FROM this:
 (TABLE this: x, db: ), expression:
 (SELECT expressions:
  (COLUMN this:
   (FUNC a:
    (COLUMN this: a, db: , table: ), b:
    (COLUMN this: b, db: , table: )), db: , table: )))

Finally generating the new SQL.

Generator(transforms={
    SpecialUDF: lambda self, e: f"SPECIAL_UDF_INVERSE({self.sql(e, 'b')}, {self.sql(e, 'a')})"
}).generate(expression)
SELECT SPECIAL_UDF_INVERSE(b, a) FROM x

Parse Errors

A syntax error will result in an parse error.

transpile("SELECT foo( FROM bar")
sqlglot.errors.ParseError: Expected )
  SELECT foo( __FROM__ bar

Unsupported Errors

Presto APPROX_DISTINCT supports the accuracy argument which is not supported in Spark.

transpile(
    'SELECT APPROX_DISTINCT(a, 0.1) FROM foo',
    read='presto',
    write='spark',
)
WARNING:root:APPROX_COUNT_DISTINCT does not support accuracy

SELECT APPROX_COUNT_DISTINCT(a) FROM foo

Rewrite Sql

Modify sql expressions like adding a CTAS

from sqlglot import Generator, parse
from sqlglot.rewriter import Rewriter

expression = parse("SELECT * FROM y")[0]
Generator().generate(Rewriter(expression).ctas('x').expression)
CREATE TABLE x AS SELECT * FROM y

Run Tests and Lint

python -m unittest && python -m pylint sqlglot/ tests/

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

sqlglot-1.0.0.tar.gz (23.3 kB view details)

Uploaded Source

Built Distribution

sqlglot-1.0.0-py3-none-any.whl (24.0 kB view details)

Uploaded Python 3

File details

Details for the file sqlglot-1.0.0.tar.gz.

File metadata

  • Download URL: sqlglot-1.0.0.tar.gz
  • Upload date:
  • Size: 23.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.6.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.61.2 CPython/3.9.6

File hashes

Hashes for sqlglot-1.0.0.tar.gz
Algorithm Hash digest
SHA256 df271a2ccc8830d7e3ca576685b7bd2993289c5bcf4f500a8e75ba2024d13222
MD5 9dc175a1f8e0da7da742327cd6247d2a
BLAKE2b-256 f4ce91fa4ad84c77bbc8ead34648e0fe6447ef99c1d90311395a142378f14ec3

See more details on using hashes here.

Provenance

File details

Details for the file sqlglot-1.0.0-py3-none-any.whl.

File metadata

  • Download URL: sqlglot-1.0.0-py3-none-any.whl
  • Upload date:
  • Size: 24.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.6.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.61.2 CPython/3.9.6

File hashes

Hashes for sqlglot-1.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 5aae42cc01213129460326736ad5d103cc52fe0cdf01d429b65d25da3e285a7d
MD5 acbf6207400f99ce1c3a7be8eee5277c
BLAKE2b-256 8e5014a0d4b3c1796098523d18e74e1faaf8078781649eb2d0be3a52e661c58a

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page