Skip to main content

A high-level, easy-to-deploy non-uniform Fast Fourier Transform in PyTorch.

Project description

torchkbnufft

LICENSE CI Badge

Documentation | GitHub | Notebook Examples

Simple installation from PyPI:

pip install torchkbnufft

About

torchkbnufft implements a non-uniform Fast Fourier Transform [1, 2] with Kaiser-Bessel gridding in PyTorch. The implementation is completely in Python, facilitating flexible deployment in readable code with no compilation. NUFFT functions are each wrapped as a torch.autograd.Function, allowing backpropagation through NUFFT operators for training neural networks.

This package was inspired in large part by the NUFFT implementation in the Michigan Image Reconstruction Toolbox (Matlab).

Operation Modes and Stages

The package has three major classes of NUFFT operation mode: table-based NUFFT interpolation, sparse matrix-based NUFFT interpolation, and forward/backward operators with Toeplitz-embedded FFTs [3]. Roughly, computation speed follows:

Type Speed
Toeplitz Fastest
Table Medium
Sparse Matrix Slow (not recommended)

It is generally best to start with Table interpolation and then experiment with the other modes for your problem.

Sensitivity maps can be incorporated by passing them into a KbNufft or KbNufftAdjoint object.

Documentation

An html-based documentation reference on Read the Docs.

Most files are accompanied with docstrings that can be read with help while running IPython. Example:

from torchkbnufft import KbNufft

help(KbNufft)

Examples

Jupyter notebook examples are in the notebooks/ folder.

Simple Forward NUFFT

The following code loads a Shepp-Logan phantom and computes a single radial spoke of k-space data:

import torch
import torchkbnufft as tkbn
import numpy as np
from skimage.data import shepp_logan_phantom

x = shepp_logan_phantom().astype(np.complex)
im_size = x.shape
# convert to tensor, unsqueeze batch and coil dimension
# output size: (1, 1, ny, nx)
x = torch.tensor(x).unsqueeze(0).unsqueeze(0).to(torch.complex64)

klength = 64
ktraj = np.stack(
    (np.zeros(64), np.linspace(-np.pi, np.pi, klength))
)
# convert to tensor, unsqueeze batch dimension
# output size: (2, klength)
ktraj = torch.tensor(ktraj, dtype=torch.float)

nufft_ob = tkbn.KbNufft(im_size=im_size)
# outputs a (1, 1, klength) vector of k-space data
kdata = nufft_ob(x, ktraj)

A detailed example of basic NUFFT usage is here.

SENSE-NUFFT

The package also includes utilities for working with SENSE-NUFFT operators. The above code can be modified to include sensitivity maps.

smaps = torch.rand(1, 8, 400, 400) + 1j * torch.rand(1, 8, 400, 400)
sense_data = nufft_ob(x, ktraj, smaps=smaps.to(x))

This code first multiplies by the sensitivity coils in smaps, then computes a 64-length radial spoke for each coil. All operations are broadcast across coils, which minimizes interaction with the Python interpreter, helping computation speed.

A detailed example of SENSE-NUFFT usage is here.

Sparse Matrix Precomputation

Sparse matrices are an alternative to table interpolation. Their speed can vary, but they are a bit more accurate than standard table mode. The following code calculates sparse interpolation matrices and uses them to compute a single radial spoke of k-space data:

adjnufft_ob = tkbn.KbNufftAdjoint(im_size=im_size)

# precompute the sparse interpolation matrices
interp_mats = tkbn.calc_tensor_spmatrix(
    ktraj,
    im_size=im_size
)

# use sparse matrices in adjoint
image = adjnufft_ob(kdata, ktraj, interp_mats)

Sparse matrix multiplication is only implemented for real numbers in PyTorch, which can limit their speed. A detailed example of sparse matrix precomputation usage is here.

Toeplitz Embedding

The package includes routines for calculating embedded Toeplitz kernels and using them as FFT filters for the forward/backward NUFFT operations [3]. This is very useful for gradient descent algorithms that must use the forward/backward ops in calculating the gradient. The following code shows an example:

toep_ob = tkbn.ToepNufft()

# precompute the embedded Toeplitz FFT kernel
kernel = tkbn.calc_toeplitz_kernel(ktraj, im_size)

# use FFT kernel from embedded Toeplitz matrix
image = toep_ob(image, kernel)

A detailed example of Toeplitz embedding usage is included here.

Running on the GPU

All of the examples included in this repository can be run on the GPU by sending the NUFFT object and data to the GPU prior to the function call, e.g.,

adjnufft_ob = adjnufft_ob.to(torch.device('cuda'))
kdata = kdata.to(torch.device('cuda'))
ktraj = ktraj.to(torch.device('cuda'))

image = adjnufft_ob(kdata, ktraj)

PyTorch will throw errors if the underlying dtype and device of all objects are not matching. Be sure to make sure your data and NUFFT objects are on the right device and in the right format to avoid these errors.

Computation Speed

The following computation times in seconds were observed on a workstation with a Xeon E5-2698 CPU and an Nvidia Quadro GP100 GPU for a 15-coil, 405-spoke 2D radial problem. CPU computations were limited to 8 threads and done with 64-bit floats, whereas GPU computations were done with 32-bit floats. The benchmark used torchkbnufft version 1.0.0 and torch version 1.7.1.

(n) = normal, (spm) = sparse matrix, (toep) = Toeplitz embedding, (f/b) = forward/backward combined

Operation CPU (n) CPU (spm) CPU (toep) GPU (n) GPU (spm) GPU (toep)
Forward NUFFT 0.82 0.77 0.058 (f/b) 2.58e-02 7.44e-02 3.03e-03 (f/b)
Adjoint NUFFT 0.75 0.76 N/A 3.56e-02 7.93e-02 N/A

Profiling for your machine can be done by running

pip install -r dev-requirements.txt
python profile_torchkbnufft.py

Other Packages

For users interested in NUFFT implementations for other computing platforms, the following is a partial list of other projects:

  1. TF KB-NUFFT (KB-NUFFT for TensorFlow)
  2. SigPy (for Numpy arrays, Numba (for CPU) and CuPy (for GPU) backends)
  3. FINUFFT (for MATLAB, Python, Julia, C, etc., very efficient)
  4. NFFT (for Julia)
  5. PyNUFFT (for Numpy, also has PyCUDA/PyOpenCL backends)

References

  1. Fessler, J. A., & Sutton, B. P. (2003). Nonuniform fast Fourier transforms using min-max interpolation. IEEE Transactions on Signal Processing, 51(2), 560-574.

  2. Beatty, P. J., Nishimura, D. G., & Pauly, J. M. (2005). Rapid gridding reconstruction with a minimal oversampling ratio. IEEE Transactions on Medical Imaging, 24(6), 799-808.

  3. Feichtinger, H. G., Gr, K., & Strohmer, T. (1995). Efficient numerical methods in non-uniform sampling theory. Numerische Mathematik, 69(4), 423-440.

Citation

If you use the package, please cite:

@conference{muckley:20:tah,
  author = {M. J. Muckley and R. Stern and T. Murrell and F. Knoll},
  title = {{TorchKbNufft}: A High-Level, Hardware-Agnostic Non-Uniform Fast {Fourier} Transform},
  booktitle = {ISMRM Workshop on Data Sampling \& Image Reconstruction},
  year = 2020
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

torchkbnufft-1.2.0.post1.tar.gz (31.5 kB view details)

Uploaded Source

Built Distribution

torchkbnufft-1.2.0.post1-py3-none-any.whl (36.7 kB view details)

Uploaded Python 3

File details

Details for the file torchkbnufft-1.2.0.post1.tar.gz.

File metadata

  • Download URL: torchkbnufft-1.2.0.post1.tar.gz
  • Upload date:
  • Size: 31.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/4.0.1 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.61.0 CPython/3.9.5

File hashes

Hashes for torchkbnufft-1.2.0.post1.tar.gz
Algorithm Hash digest
SHA256 ce50ad80e2d942308ec526d1e99e29017e9bad2172562a8a380dfa5d6d477241
MD5 d2b9e805000333a150e5fd80ef6335c4
BLAKE2b-256 84fdc5fd441e4c9ca2b0d63701a054fdf76aff16f25e6078235e6b8f1e02c768

See more details on using hashes here.

File details

Details for the file torchkbnufft-1.2.0.post1-py3-none-any.whl.

File metadata

  • Download URL: torchkbnufft-1.2.0.post1-py3-none-any.whl
  • Upload date:
  • Size: 36.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/4.0.1 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.61.0 CPython/3.9.5

File hashes

Hashes for torchkbnufft-1.2.0.post1-py3-none-any.whl
Algorithm Hash digest
SHA256 d455218545a5d4b0ac85ea02932e71623be2eb6aa44b32b4b9c7f07ddf2e26fb
MD5 3c6f70dcb6a359a51a90f770b212d9d5
BLAKE2b-256 24b7e7981af8b15bce6bc15e91057bb4fd2856c32792c71607b2d1fbf18fc0a8

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page