Skip to main content

Copy data from PostgreSQL databases to MySQL

Project description

# Description
`psql2mysql` provides some tooling for migrating data from PostgreSQL to
MariaDB. It was main target for this tool is to migrate the databases of
an [OpenStack](https://www.openstack.org) deployment from PostgreSQL to
MariaDB. But it should work for other databases as well.

The tool itself is based on SQLalchemy and it's builtin type abstraction to
handle some of the differences between both database.

Currently its focused on moving from PostgreSQL to MariaDB. Other MySQL like
databases haven't been tested yet. Neither is the reverse direction of the
migration (MariaDB to PostgreSQL) implemented. Both of these features should
be easy to add however.

# Usage

## Installation

```
mkdir venv
virtualenv venv
. venv/bin/activate
pip install -r requirements.txt
python setup.py install
```

## Prerequisites
* Before using `psql2mysql` for migration, the destination database and all
its tables need to exist already.
* `psql2mysql` does not do any checks currently if the schema of the target
and destination database are compatible. In the case of a migration of an
OpenStack database it is assumed that the respective db_sync tool (e.g.
`keystone-manage db_sync`) of the OpenStack service has been called before
running `psql2mysql`.
* The identity provided for the source database needs to have enough
privileges to read all data and schema in the specified database.
* Additionally to write access, the identity provided for the target database
needs privileges to temporary disable constraints and foreign keys for the
duration of the migration.

## Commands
`psql2mysql` currenty provides two subcommands:

* `precheck`

Runs a couple of checks on the tables of the source database. Currently
only searches all text columns for characters that are not compatible with
the "utf8" encoding of MariaDB. Prints out the rows (and affected columns)
that contains incompatible strings.

* `migrate`

Runs the acutal migration. Will go through, the database table by table
and migrate all rows to the target database.

* `purge-tables`

Purges all tables in the target database. Tables with names related to
alembic and SQLalchemy-migrate are skipped and not purged. This subcommand
can be executed optionally after a failed migration attempt. The `migrate`
subcommand will also purge all tables in the target database in the same
way before it starts to copy any data from the source database. But as
that operation can take a significant amount of time it might make sense
to run the `purge-tables` subcommand separately before `migrate` in order
to reduce the overall downtime of the source database.

### Options

* `exclude-deleted`

When migrating the data from source to target database, exclude soft-deleted
rows. Set to true by default.

* `batch`

Process more databases in one run of `psql2mysql`. The argument for this option
is a path to YML file that lists set of databases with their source and target
connection strings. The format of YML file is:

```
database1:
source: source URI for database1
target: target URI for database1
database2:
source: source URI for database2
...
```

* `chunk-size`

By default table migrations are broken into chunks, a number of rows at a
time. The default size is `10000`. This option with change the chunk size.

Breaking the migration into chunks improves RAM usage and allows the
migration tool to work with larger databases.

Setting `chunk-size` to `0` will disable chunking and the migration will
migrate the whole table in one go. Be warned that this will use more RAM.

## Examples

To check that it is actually possible to migrate run the `precheck`
subcommand:

```psql2mysql --source postgresql://neutron:secret@192.168.1.1/neutron precheck```

To check all databases in the batch file:

```psql2mysql --batch databases.yml precheck```


To do the actual migraton:
```
psql2mysql \
--source postgresql://neutron:secret@192.168.1.1/neutron \
--target mysql+pymysql://neutron:evenmoresecret@192.168.1.2/neutron?charset=utf8 \
migrate
```

To migrate all databases in the batch file:

```psql2mysql --batch databases.yml migrate```


# Testing
`psql2mysql` provides a test suite. To run the tests use:

```nosetests```



Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

psql2mysql-0.5.1.tar.gz (16.1 kB view details)

Uploaded Source

Built Distribution

psql2mysql-0.5.1-py2-none-any.whl (17.6 kB view details)

Uploaded Python 2

File details

Details for the file psql2mysql-0.5.1.tar.gz.

File metadata

  • Download URL: psql2mysql-0.5.1.tar.gz
  • Upload date:
  • Size: 16.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.12.1 pkginfo/1.4.2 requests/2.19.1 setuptools/40.4.3 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/2.7.15rc1

File hashes

Hashes for psql2mysql-0.5.1.tar.gz
Algorithm Hash digest
SHA256 28d60bb64b18c765831732b586e35260e48e8577b4fa28be469c725d4a858e61
MD5 ac8a0a113ea831c7549f55ec0def4b01
BLAKE2b-256 5f12dc2bde26dbc5658ffc4f6de8a60226dc035bf362f8fd0bf47c3d2c07427c

See more details on using hashes here.

File details

Details for the file psql2mysql-0.5.1-py2-none-any.whl.

File metadata

  • Download URL: psql2mysql-0.5.1-py2-none-any.whl
  • Upload date:
  • Size: 17.6 kB
  • Tags: Python 2
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.12.1 pkginfo/1.4.2 requests/2.19.1 setuptools/40.4.3 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/2.7.15rc1

File hashes

Hashes for psql2mysql-0.5.1-py2-none-any.whl
Algorithm Hash digest
SHA256 fd0f444c78c974995eb64a6d2d04db070cd451a13f0348cd04cfc10e2dfa1133
MD5 2d7b42776d89cb178d21b43afea4f5ce
BLAKE2b-256 41deb3d118050c1a91623b0f900ea8f7eb00774caef6094b3eb8642376c1940b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page