deep learning
Paper summary: “Perceiver : General Perception with Iterative Attention”
arxiv: https://arxiv.org/abs/2103.03206 key points use learnable latent array with a fixed size as query in transformer module in perceiver architecture raw inputs can be long since it will be use as key and value in transformer module do cross attention to fusion input data into latent array. this allows perceiver to Read more…