The post discusses the application of block sparsity on MLP modules in vision transformers, showing promising speedups with minimal accuracy drop. It explains the training and inference steps, provides microbenchmarking results, and showcases the speedup and accuracy achieved on a specific ViT model. Future steps and potential optimizations for block sparsity on vision transformers are also mentioned.

6m read timeFrom pytorch.org
Post cover image
Table of contents
IntroductionApproachResults: MicrobenchmarksResults: Vision TransformersNext Steps

Sort: