Industry-strength Natural Language Processing extensions for Keras.
Project description
KerasNLP
KerasNLP is a simple and powerful API for building Natural Language Processing (NLP) models within the Keras ecosystem.
KerasNLP provides modular building blocks following standard Keras interfaces (layers, metrics) that allow you to quickly and flexibly iterate on your task. Engineers working in applied NLP can leverage the library to assemble training and inference pipelines that are both state-of-the-art and production-grade.
KerasNLP can be understood as a horizontal extension of the Keras API — components are first-party Keras objects that are too specialized to be added to core Keras, but that receive the same level of polish as the rest of the Keras API.
We are a new and growing project, and welcome contributions.
Quick Links
For everyone
For contributors
Quick Start
Install the latest release:
pip install keras-nlp --upgrade
Tokenize text, build a tiny transformer, and train a single batch:
import keras_nlp
import tensorflow as tf
from tensorflow import keras
# Tokenize some inputs with a binary label.
vocab = ["[UNK]", "the", "qu", "##ick", "br", "##own", "fox", "."]
sentences = ["The quick brown fox jumped.", "The fox slept."]
tokenizer = keras_nlp.tokenizers.WordPieceTokenizer(
vocabulary=vocab,
sequence_length=10,
)
x, y = tokenizer(sentences), tf.constant([1, 0])
# Create a tiny transformer.
inputs = keras.Input(shape=(None,), dtype="int32")
outputs = keras_nlp.layers.TokenAndPositionEmbedding(
vocabulary_size=len(vocab),
sequence_length=10,
embedding_dim=16,
)(inputs)
outputs = keras_nlp.layers.TransformerEncoder(
num_heads=4,
intermediate_dim=32,
)(outputs)
outputs = keras.layers.GlobalAveragePooling1D()(outputs)
outputs = keras.layers.Dense(1, activation="sigmoid")(outputs)
model = keras.Model(inputs, outputs)
# Run a single batch of gradient descent.
model.compile(optimizer="adam", loss="binary_crossentropy", jit_compile=True)
model.train_on_batch(x, y)
Compatibility
We follow Semantic Versioning, and plan to
provide backwards compatibility guarantees both for code and saved models built
with our components. While we continue with pre-release 0.y.z
development, we
may break compatibility at any time and APIs should not be consider stable.
Citing KerasNLP
If KerasNLP helps your research, we appreciate your citations. Here is the BibTeX entry:
@misc{kerasnlp2022,
title={KerasNLP},
author={Watson, Matthew, and Qian, Chen, and Zhu, Scott and Chollet, Fran\c{c}ois and others},
year={2022},
howpublished={\url{https://github.com/keras-team/keras-nlp}},
}
Thank you to all of our wonderful contributors!
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file keras-nlp-0.3.0.tar.gz
.
File metadata
- Download URL: keras-nlp-0.3.0.tar.gz
- Upload date:
- Size: 85.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.13
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | b7d086e940adeaa3829994e13ea20614054b335f5d138d842ab8b011bbebef0d |
|
MD5 | c56e655b198af328b093d8629252d3ac |
|
BLAKE2b-256 | 260364c0e41333ef050a33bea8ca2cb645b5dc4eb36c0486790f935329468218 |
Provenance
File details
Details for the file keras_nlp-0.3.0-py3-none-any.whl
.
File metadata
- Download URL: keras_nlp-0.3.0-py3-none-any.whl
- Upload date:
- Size: 142.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.13
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5419ea5dc16a8505a43a4911faea18937e76ae8b1aabeb7dc4515c43d35a1ba4 |
|
MD5 | 4e03e12dcef260dcd4f9d62abf8c0f94 |
|
BLAKE2b-256 | 05ebd1a84fc539f806eba0a41491a6b896a80719603cd2821f7802098f3345f0 |