Improve scene segmentation with smaller models by distilling knowledge with SKR+PEA

--

Improve scene segmentation with smaller models by distilling knowledge with SKR+PEA

Transformer-based Knowledge Distillation for Efficient Semantic Segmentation of Road-driving Scenes
arXiv paper abstract https://arxiv.org/abs/2202.13393v1
arXiv PDF paper https://arxiv.org/pdf/2202.13393v1.pdf

For scene understanding in robotics and automated driving, there is … interest in solving semantic segmentation tasks with transformer-based methods.

However, effective transformers are always too cumbersome and computationally expensive to solve semantic segmentation in real time

… Knowledge Distillation (KD) speeds up inference and maintains accuracy while transferring knowledge from a pre-trained cumbersome teacher model to a compact student model.

… present a novel KD framework … training compact transformers by transferring the knowledge from feature maps and patch embeddings of large transformers.

… two modules … proposed: (1) … an efficient relation-based KD framework … (SKR); (2) … (PEA) … performs the dimensional transformation of patch embeddings. The combined KD framework is called SKR+PEA.

… approach outperforms recent state-of-the-art KD frameworks and rivals the time-consuming pre-training method. …

Stay up to date. Subscribe to my posts https://morrislee1234.wixsite.com/website/contact
Web site with my other posts by category https://morrislee1234.wixsite.com/website

LinkedIn https://www.linkedin.com/in/morris-lee-47877b7b

Photo by Yulia Matvienko on Unsplash

--

--

AI News Clips by Morris Lee: News to help your R&D
AI News Clips by Morris Lee: News to help your R&D

Written by AI News Clips by Morris Lee: News to help your R&D

A computer vision consultant in artificial intelligence and related hitech technologies 37+ years. Am innovator with 66+ patents and ready to help a firm's R&D.

No responses yet