MLP-Mixer: An all-MLP Architecture for Vision

 Why:

Comparable results can be achieved for vision related tasks without using CNN or ViT (Self-attention) simply by using MLP

How:

Like Vision transformers image patches are fed as input tokens and it process through MLP’s

What:

This technique is called as MLP mixer , it has 2 kind of MLP’s , one which interacts with channels and other interacts with spatial region

TL: DR

Architecture is self-explanatory, main idea here is we have 2 type of MLP layers ,

MLP1 deals with it’s channel component (Own Patch) and MLP2 interacts across spatial region (Other patches)

Points to Ponder:

1. Still property of parameter sharing isn’t used completely as much we do in CNN

2. Like vision transformers interesting to observe what happens if we change the order of patches or overlapping patch?

link to the paper : https://arxiv.org/pdf/2105.01601.pdf

Comments

Popular posts from this blog

NeurIPS 2021 - Curated papers - Part 2

DALL-E (Zero-Shot Text-to-Image Generation) -PART(2/2)

An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale