This Python package implements Perceiver: General Perception with Iterative Attention by Andrew Jaegle in TensorFlow. This model builds on top of Transformers such that the data only enters through the cross attention mechanism (see figure) and allow it to scale to hundreds of thousands of inputs, like ConvNets. This, in part also solves the Transformers Quadratic compute and memory bottleneck.
Run the following to install:
pip install perceiver
perceiver, along with tools you need to develop and test, run the following in your virtualenv:
git clone https://github.com/Rishit-dagli/Perceiver.git # or clone your own fork cd perceiver pip install -e .[dev]
A bit about Perceiver
The Perceiver model aims to deal with arbitrary configurations of different modalities using a single transformer-based architecture. Transformers are often flexible and make few assumptions about their inputs, but that also scale quadratically with the number of inputs in terms of both memory and computation. This model proposes a mechanism that makes it possible to deal with high-dimensional inputs, while retaining the expressivity and flexibility to deal with arbitrary input configurations.
The idea here is to introduce a small set of latent units that forms an attention bottleneck through which the inputs must pass. This avoids the quadratic scaling problem of all-to-all attention of a classical transformer. The model can be seen as performing a fully end-to-end clustering of the inputs, with the latent units as the cluster centres, leveraging a highly asymmetric crossattention layer. For spatial information the authors compensate for the lack of explicit grid structures in our model by associating Fourier feature encodings.
from perceiver import Perceiver import tensorflow as tf model = Perceiver( input_channels = 3, # number of channels for each token of the input input_axis = 2, # number of axis for input data (2 for images, 3 for video) num_freq_bands = 6, # number of freq bands, with original value (2 * K + 1) max_freq = 10., # maximum frequency, hyperparameter depending on how fine the data is depth = 6, # depth of net num_latents = 256, # number of latents latent_dim = 512, # latent dimension cross_heads = 1, # number of heads for cross attention. paper said 1 latent_heads = 8, # number of heads for latent self attention, 8 cross_dim_head = 64, latent_dim_head = 64, num_classes = 1000, # output number of classes attn_dropout = 0., ff_dropout = 0., ) img = tf.random.normal([1, 224, 224, 3]) # replicating 1 imagenet image model(img) # (1, 1000)