Skip to main content

Python utilities for Breakthrough Listen SETI observations

Project description

Build Status Documentation Status codecov JOSS status

Breakthrough Listen I/O Methods for Python.

Filterbank + Raw file readers

This repository contains Python 2/3 readers for interacting with Sigproc filterbank (.fil), HDF5 (.h5) and guppi raw (.raw) files, as used in the Breakthrough Listen search for intelligent life.

Installation

System Dependencies

Sometimes the pip installation can fail if a system dependency is not installed. To fix this, make sure you have curl and install the required system dependencies with the command bellow:

Debian/Ubuntu
curl https://raw.githubusercontent.com/UCBerkeleySETI/blimpy/master/dependencies.txt | xargs -n 1 sudo apt install --no-install-recommends -y

Manual Installation

The latest release can be installed via pip directly from this repository:

python3 -m pip install -U git+https://github.com/UCBerkeleySETI/blimpy

Or, the latest version of the development code can be installed from the github repo and then run python setup.py install or pip install . (with sudo if required), or by using the following terminal command:

python3 -m pip install -U https://github.com/UCBerkeleySETI/blimpy/tarball/master

To install everything required to run the unit tests, run:

python3 -m pip install -e .[full]

You will need numpy, h5py, astropy, scipy, and matplotlib as dependencies. A pip install should pull in numpy, h5py, and astropy, but you may still need to install scipy and matplotlib separately. To interact with compressed files, you'll need the hdf5plugin package too.

Note that h5py generally needs to be installed in this way:

$ python3 -m pip install --no-binary=h5py h5py

Command line utilities

After installation, the following command will display the metadata (header) values and some information about the data matrix:

watutil -i

Other command line utilities available post-installation:

  • bldice, Extract a smaller frequency region from a Filterbank file (.fil or .h5 file).
  • calcload, Calculate the Waterfall max_load value needed to load the entire data array for a given Filterbank file.
  • dsamp, Down-sample (time-dimension only) from one Filterbank file to another.
  • fil2h5, Convert a .fil file into .h5 format.
  • h52fil, Convert an .h5 file into .fil format.
  • matchfils, Check if two .fil files are the same.
  • peek, Display a selected portion of values from the data matrix of a Filterbank file.
  • rawhdr, Display the header fields of a raw guppi file.
  • rawutil, Plot data in a guppi raw file.
  • srcname, Patch the header source_name field in a .h5 file.
  • stax, For a collection of .h5 or .fil files sharing the same frequency range, create a vertical stack of waterfall plots as a single PNG file.
  • stix, For a single very large Filterbank file, create a horizontal or vertical stack of waterfall plots as a single PNG file.
  • watutil, Information/read/write/plot utility for Filterbank files.

Use the -h flag to any of the above command line utilities to display their available arguments.

Reading blimpy filterbank files in .fil or .h5 format

The blimpy.Waterfall provides a Python API for interacting with filterbank data. It supports all BL filterbank data products; see this example Jupyter notebook for an overview.

From the python, ipython or jupiter notebook environments.

from blimpy import Waterfall
fb = Waterfall('/path/to/filterbank.fil')
#fb = Waterfall('/path/to/filterbank.h5') #works the same way
fb.info()
data = fb.data

Reading guppi raw files

The Guppi Raw format can be read using the GuppiRaw class from guppi.py:

from blimpy import GuppiRaw
gr = GuppiRaw('/path/to/guppirawfile.raw')

header, data = gr.read_next_data_block()

or

from blimpy import GuppiRaw
gr = GuppiRaw('/path/to/guppirawfile.raw')

for header, data_x, data_y in gr.get_data():
    # process data

Note: most users should start analysis with filterbank files, which are smaller in size and have been generated from the guppi raw files.

Using blimpy inside Docker

The blimpy images are pushed to a public repository after each successful build on Travis. If you have Docker installed, you can run the following commands to pull our images, which have the environment and dependencies set up for you.

docker pull fx196/blimpy:py3_kern_stable

Here is a more complete guide on using blimpy in Docker.

Further reading

A detailed overview of the data formats used in Breakthrough Listen can be found in our data format paper. An archive of data files from the Breakthrough Listen program is provided at seti.berkeley.edu/opendata.

If you have any requests or questions, please lets us know!

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

blimpy-2.1.4.tar.gz (89.3 kB view details)

Uploaded Source

Built Distribution

blimpy-2.1.4-py3-none-any.whl (104.3 kB view details)

Uploaded Python 3

File details

Details for the file blimpy-2.1.4.tar.gz.

File metadata

  • Download URL: blimpy-2.1.4.tar.gz
  • Upload date:
  • Size: 89.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.11.0

File hashes

Hashes for blimpy-2.1.4.tar.gz
Algorithm Hash digest
SHA256 5ad923ed50bae2f421ba3024fd03a87046f004789e19b8d5e2fcb6f78c3a6dae
MD5 f3aeb510fd6d319fa8283f09c06e76cc
BLAKE2b-256 d20d7ffc61a69a8582c810057c07445ed23d1617defe462ab9059981919bc2fc

See more details on using hashes here.

File details

Details for the file blimpy-2.1.4-py3-none-any.whl.

File metadata

  • Download URL: blimpy-2.1.4-py3-none-any.whl
  • Upload date:
  • Size: 104.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.11.0

File hashes

Hashes for blimpy-2.1.4-py3-none-any.whl
Algorithm Hash digest
SHA256 213425ea2fcdc2ed74dc4074b894b2d2861bb3e197308cdeb3707f6301a7ba0a
MD5 6dea757b79d06572b4778bf8769f0eb9
BLAKE2b-256 373b3f315b189e1f25db9682f42a9227a396957d0bdd11623989ed5416de81aa

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page