Package for training Gaussian process-like Bayesian Neural Networks with composite structure.
Project description
AutoBNN
This library contains code to specify BNNs that correspond to various useful GP kernels and assemble them into models using operators such as Addition, Multiplication and Changepoint.
It is based on the ideas in the following papers:
-
Lassi Meronen, Martin Trapp, Arno Solin. Periodic Activation Functions Induce Stationarity. NeurIPS 2021.
-
Tim Pearce, Russell Tsuchida, Mohamed Zaki, Alexandra Brintrup, Andy Neely. Expressive Priors in Bayesian Neural Networks: Kernel Combinations and Periodic Functions. UAI 2019.
-
Feras A. Saad, Brian J. Patton, Matthew D. Hoffman, Rif A. Saurous, Vikash K. Mansinghka. Sequential Monte Carlo Learning for Time Series Structure Discovery. ICML 2023.
Setup
AutoBNN has three additional dependencies beyond those used by the core
Tensorflow Probability package: flax, scipy and jaxtyping. These
can be installed by running setup_autobnn.sh
.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file autobnn-0.0.1.dev0.tar.gz
.
File metadata
- Download URL: autobnn-0.0.1.dev0.tar.gz
- Upload date:
- Size: 29.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.11.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f5d8b25232bc0675b361f5db79c160081855768f97cbf22a6cac957de8c5a4e6 |
|
MD5 | cc4c19740fe76b921d087769853d037f |
|
BLAKE2b-256 | 9e65acfc5080494e174e25e00924bfadd09395ef1ffe0c80b9243d853f7266e0 |