Skip to main content

Microsoft Health Futures package to work with multi-modal health data

Project description

HI-ML Multimodal Toolbox

This toolbox provides models for working with multi-modal health data. The code is available on GitHub and Hugging Face 🤗.

Getting started

The best way to get started is by running the phrase grounding notebook. All the dependencies will be installed upon execution, so Python 3.7 and Jupyter are the only requirements to get started.

The notebook can also be run on Binder, without the need to download any code or install any libraries:

Binder

Installation

The latest version can be installed using pip:

pip install "git+https://github.com/microsoft/hi-ml.git#subdirectory=hi-ml-multimodal"

Development

For development, it is recommended to clone the repository and set up the environment using conda:

git clone https://github.com/microsoft/hi-ml.git
cd hi-ml-multimodal
make env

This will create a conda environment named multimodal and install all the dependencies to run and test the package.

You can visit the API documentation for a deeper understanding of our tools.

Hugging Face 🤗

While the GitHub repository provides examples and pipelines to use our models, the weights and model cards are hosted on Hugging Face 🤗.

Credit

If you use our code or models in your research, please cite the manuscript (accepted to be presented at the European Conference on Computer Vision (ECCV) 2022).

APA

Boecking, B., Usuyama, N., Bannur, S., Castro, D., Schwaighofer, A., Hyland, S., Wetscherek, M., Naumann, T., Nori, A., Alvarez-Valle, J., Poon, H., & Oktay, O. (2022). Making the Most of Text Semantics to Improve Biomedical Vision–Language Processing (preprint)

BibTeX

@misc{https://doi.org/10.48550/arxiv.2204.09817,
  doi = {10.48550/ARXIV.2204.09817},
  url = {https://arxiv.org/abs/2204.09817},
  author = {Boecking, Benedikt and Usuyama, Naoto and Bannur, Shruthi and Castro, Daniel C. and Schwaighofer, Anton and Hyland, Stephanie and Wetscherek, Maria and Naumann, Tristan and Nori, Aditya and Alvarez-Valle, Javier and Poon, Hoifung and Oktay, Ozan},
  keywords = {Computer Vision and Pattern Recognition (cs.CV), Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
  title = {Making the Most of Text Semantics to Improve Biomedical Vision-Language Processing},
  publisher = {arXiv},
  year = {2022},
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

hi-ml-multimodal-0.1.0.tar.gz (18.6 kB view details)

Uploaded Source

Built Distribution

hi_ml_multimodal-0.1.0-py3-none-any.whl (26.5 kB view details)

Uploaded Python 3

File details

Details for the file hi-ml-multimodal-0.1.0.tar.gz.

File metadata

  • Download URL: hi-ml-multimodal-0.1.0.tar.gz
  • Upload date:
  • Size: 18.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.7.13

File hashes

Hashes for hi-ml-multimodal-0.1.0.tar.gz
Algorithm Hash digest
SHA256 a16757f8750814e14e8c4e2dfe6e081a0705f57b231a1bf0a352dd769137061a
MD5 70070f5a2ec208223bf28b2ffef4007e
BLAKE2b-256 eb615fbff73732c108036688442ce6265c29f12970dd1d7833a98254fa865814

See more details on using hashes here.

File details

Details for the file hi_ml_multimodal-0.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for hi_ml_multimodal-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 b274c7c5026691c22e0a8c6a00b4eeea6f2ee8af861a055978aba60c4cffa190
MD5 3d20a19c1842a2858bf1452607ee4fc3
BLAKE2b-256 6562b44e87e4ecfc674f3e25e11b2d7fa93b5f3a7ea2b527e42ececf0ad1207e

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page