Skip to main content

Facebook AI Research Sequence-to-Sequence Toolkit

Project description



Support Ukraine MIT License Latest Release Build Status Documentation Status CicleCI Status


Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks.

We provide reference implementations of various sequence modeling papers:

List of implemented papers

What's New:

Previous updates

Features:

We also provide pre-trained models for translation and language modeling with a convenient torch.hub interface:

en2de = torch.hub.load('pytorch/fairseq', 'transformer.wmt19.en-de.single_model')
en2de.translate('Hello world', beam=5)
# 'Hallo Welt'

See the PyTorch Hub tutorials for translation and RoBERTa for more examples.

Requirements and Installation

  • PyTorch version >= 1.5.0
  • Python version >= 3.6
  • For training new models, you'll also need an NVIDIA GPU and NCCL
  • To install fairseq and develop locally:
git clone https://github.com/pytorch/fairseq
cd fairseq
pip install --editable ./

# on MacOS:
# CFLAGS="-stdlib=libc++" pip install --editable ./

# to install the latest stable release (0.10.x)
# pip install fairseq
  • For faster training install NVIDIA's apex library:
git clone https://github.com/NVIDIA/apex
cd apex
pip install -v --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" \
  --global-option="--deprecated_fused_adam" --global-option="--xentropy" \
  --global-option="--fast_multihead_attn" ./
  • For large datasets install PyArrow: pip install pyarrow
  • If you use Docker make sure to increase the shared memory size either with --ipc=host or --shm-size as command line options to nvidia-docker run .

Getting Started

The full documentation contains instructions for getting started, training new models and extending fairseq with new model types and tasks.

Pre-trained models and examples

We provide pre-trained models and pre-processed, binarized test sets for several tasks listed below, as well as example training and evaluation commands.

We also have more detailed READMEs to reproduce results from specific papers:

Join the fairseq community

License

fairseq(-py) is MIT-licensed. The license applies to the pre-trained models as well.

Citation

Please cite as:

@inproceedings{ott2019fairseq,
  title = {fairseq: A Fast, Extensible Toolkit for Sequence Modeling},
  author = {Myle Ott and Sergey Edunov and Alexei Baevski and Angela Fan and Sam Gross and Nathan Ng and David Grangier and Michael Auli},
  booktitle = {Proceedings of NAACL-HLT 2019: Demonstrations},
  year = {2019},
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fairseq-0.11.1.tar.gz (9.6 MB view details)

Uploaded Source

Built Distributions

fairseq-0.11.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl (11.0 MB view details)

Uploaded CPython 3.8 manylinux: glibc 2.5+ x86-64

fairseq-0.11.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl (10.9 MB view details)

Uploaded CPython 3.7m manylinux: glibc 2.5+ x86-64

fairseq-0.11.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl (10.9 MB view details)

Uploaded CPython 3.6m manylinux: glibc 2.5+ x86-64

File details

Details for the file fairseq-0.11.1.tar.gz.

File metadata

  • Download URL: fairseq-0.11.1.tar.gz
  • Upload date:
  • Size: 9.6 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.8.12

File hashes

Hashes for fairseq-0.11.1.tar.gz
Algorithm Hash digest
SHA256 1a84cde1ae1f4632bb1a714669a08b02625d38198d693a03ab45788ab049861f
MD5 38e25c209e8f6c0d5816b5a4889bcf70
BLAKE2b-256 2b7c44b138514088b440ad8b4269f3feb3e1216208a7f46143fd51eb577166de

See more details on using hashes here.

Provenance

File details

Details for the file fairseq-0.11.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl.

File metadata

File hashes

Hashes for fairseq-0.11.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 ea1b91a4f2875f7c857fb1f2b579debb3101d003833397782e4552c47c24b40b
MD5 84fa4c9788b866003129876e7f8f7246
BLAKE2b-256 fecb1ae7064955c835f383c72403524c12ae1abfbab86b344a30e3606d9c62d3

See more details on using hashes here.

Provenance

File details

Details for the file fairseq-0.11.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl.

File metadata

File hashes

Hashes for fairseq-0.11.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 9d9403c3e1eb9b11585f591b37ed7f79ba527ca59ebf2aa1de644a4ac4684a6e
MD5 63ba4965c9f96b77dedad4e9ec989808
BLAKE2b-256 06ec854ccc33367d73ac0b9e3c0de10cf6f7867f533871840a89e0fe23489bc3

See more details on using hashes here.

Provenance

File details

Details for the file fairseq-0.11.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl.

File metadata

File hashes

Hashes for fairseq-0.11.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 db8fb37165c6999621a45dcfb81deb1159c9d9806bf6d9c841e840819c2623bf
MD5 5fa6fca1e2fca69a25dc89984eff9fab
BLAKE2b-256 20a04442145278e1841f13da49cf553ab198278d3fd6f535003645753ccc180f

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page