Загрузка страницы

tinyML Research Symposium: AugViT: Improving Vision Transformer Training by Marrying Attention and..

Zhongzhi YU, PhD Student, EIC Lab at Georgia Institute of Technology
AugViT: Improving Vision Transformer Training by Marrying Attention and Data Augmentation
Despite the impressive accuracy of large-scale vision transformers (ViTs) across various tasks, it remains a challenge for small-scale ViTs (e.g., 1G inference floating points operations (FLOPs) as in LeViT) to significantly outperform state-of-the-art convolution neural networks (CNNs) in terms of the accuracy-efficiency trade-off, limiting their wider application, especially on resource-constrained devices. As analyzed in recent works, selecting an effective data augmentation technique can non-trivially improve the accuracy of small-scale ViTs. However, whether existing mainstream data augmentation techniques dedicated to CNNs are optimal for ViTs is still an open question. To this end, we propose a data augmentation framework called AugViT, which is dedicated to incorporating the key component in ViTs, i.e., self-attention, into data augmentation intensity to enable ViT’s outstanding performance across various devices. Specifically, motivated by ViT’s patch-based processing pipeline, our proposed AugViT integrates (1) a dedicated scheme for mapping the attention map in ViTs to the suggested augmentation intensity for each patch, (2) a simple but effective strategy of selecting the most effective attention map within ViTs to guide the aforementioned attention-aware data argumentation, and (3) a set of patch-level augmentation techniques that matches the patch-aware processing pipeline and enables the varying of augmentation intensities in each patch. Extensive experiments and ablation studies on two datasets and ten representative ViT models validate AugViT’s effectiveness in boosting ViTs’ performance, especially for small-scale ViTs, e.g., improving LeViT-128S’s accuracy from 76.6% to 77.1%, achieving a comparable accuracy to EfficientNet-B0 with 21.8% fewer inference FLOPs overhead on ImageNet dataset).

Видео tinyML Research Symposium: AugViT: Improving Vision Transformer Training by Marrying Attention and.. канала The tinyML Foundation
Показать
Комментарии отсутствуют
Введите заголовок:

Введите адрес ссылки:

Введите адрес видео с YouTube:

Зарегистрируйтесь или войдите с
Информация о видео
27 мая 2023 г. 19:43:23
00:14:19
Другие видео канала
tinyML Talks Taiwan in Mandarin and English: Discovering tinyMLtinyML Talks Taiwan in Mandarin and English: Discovering tinyMLtinyML Summit 2022: Sensing Applications as a Driver for TinyML SolutionstinyML Summit 2022: Sensing Applications as a Driver for TinyML SolutionstinyML Neuromorphic Engineering Forum - Sensors SessiontinyML Neuromorphic Engineering Forum - Sensors SessiontinyML Vision Challenge - Himax & Edge ImpulsetinyML Vision Challenge - Himax & Edge ImpulsetinyML Talks Chao Xu: Enabling Neural network at the low power edge: A neural network compiler...tinyML Talks Chao Xu: Enabling Neural network at the low power edge: A neural network compiler...SensMACH 2020 Daniel Situnayake: Embedded machine learning in the real worldSensMACH 2020 Daniel Situnayake: Embedded machine learning in the real worldtinyML Talks: Empowering the Edge: Practical Applications of Embedded Machine Learning on MCUstinyML Talks: Empowering the Edge: Practical Applications of Embedded Machine Learning on MCUstinyML Talks: Efficient AI for Wildlife ConservationtinyML Talks: Efficient AI for Wildlife ConservationtinyML Research Symposium 2022: Towards Agile Design of Neural Processing Units with ChiseltinyML Research Symposium 2022: Towards Agile Design of Neural Processing Units with ChiseltinyML Talks Phoenix: Novel Device and Materials in Emerging Memory for Neuromorphic ComputingtinyML Talks Phoenix: Novel Device and Materials in Emerging Memory for Neuromorphic ComputingtinyML Talks - Phoenix meetup: Analog TinyML for health management using intelligent wearablestinyML Talks - Phoenix meetup: Analog TinyML for health management using intelligent wearablestinyML Talks India: Single Lead ECG Classification On Wearable and Implantable DevicestinyML Talks India: Single Lead ECG Classification On Wearable and Implantable DevicestinyML Summit 2023:Personal Computing devices use-case and applications enabled by Smart SensorstinyML Summit 2023:Personal Computing devices use-case and applications enabled by Smart SensorstinyML Talks: From the lab to the edge: Post-Training CompressiontinyML Talks: From the lab to the edge: Post-Training CompressiontinyML Talks: State of Hardware & Software Ecosystem for Low-Power ML Applications on RISC-VtinyML Talks: State of Hardware & Software Ecosystem for Low-Power ML Applications on RISC-VtinyML Talks: Meetup Italy with small-medium industriestinyML Talks: Meetup Italy with small-medium industriestinyML Hackathon Challenge  2023 -  Infineon XENSIV 60GHz Radar Sensor and devkit explanationtinyML Hackathon Challenge 2023 - Infineon XENSIV 60GHz Radar Sensor and devkit explanationtinyML Auto ML Tutorial with QeexotinyML Auto ML Tutorial with QeexotinyML On Device Learning Forum - Warren Gross: On-Device Learning For Natural Language Processing..tinyML On Device Learning Forum - Warren Gross: On-Device Learning For Natural Language Processing..EMEA 2021 tiny Talks: Building Heterogeneous TinyML PipelinesEMEA 2021 tiny Talks: Building Heterogeneous TinyML PipelinestinyML EMEA 2022- Eran Treister: Wavelet Feature Maps Compression for Image-to-Image CNNstinyML EMEA 2022- Eran Treister: Wavelet Feature Maps Compression for Image-to-Image CNNs
Яндекс.Метрика