Skip to content

Latest commit

 

History

History
20 lines (16 loc) · 790 Bytes

README.md

File metadata and controls

20 lines (16 loc) · 790 Bytes

3D-MaxViT-pytorch

Extending MaxViT (Multi-Axis Vision Transformer) to 3D Space

MaxViT

MaxViT:Multi-Axis Vision Transformer. In this paper author introduce an efficient and scalable attention model , call multi-axis attention, which consists of two aspects: blocked local and dilated global attention.


max-sa

3D MaxViT

We extend MaxViT to 3D space to accommodate medical images such as CT and MRI. We implemented 3D Max-SA(3D Multi-axis self-attention) as shown in the figure to facilitate insertion into existing medical networks such as 3DUnet.


3d-max-sa

Acknowledgement

We refer to part of the code of vit-pytorch/maxvit