Show simple item record

dc.contributor.authorLe, Duy Dung
dc.contributor.authorTran, Viet Anh
dc.contributor.authorNguyen, M. Tan
dc.contributor.authorNguyen, Tam
dc.contributor.authorNguyen, Duy Khuong
dc.contributor.authorBaraniuk, Richard G.
dc.contributor.authorHo, Nhat
dc.contributor.authorOsher, Stanley J.
dc.date.accessioned2024-10-24T08:44:56Z
dc.date.available2024-10-24T08:44:56Z
dc.date.issued2022
dc.identifier.urihttps://vinspace.edu.vn/handle/VIN/297
dc.description.abstractMulti-head attention is a driving force behind state-of-the-art transformers, which achieve remarkable performance across a variety of natural language processing (NLP) and computer vision tasks. It has been observed that for many applications, those attention heads learn redundant embedding, and most of them can be removed without degrading the performance of the model. Inspired by this observation, we propose Transformer with a Mixture of Gaussian Keys (Transformer-MGK), a novel transformer architecture that replaces redundant heads in transformers with a mixture of keys at each head. These mixtures of keys follow a Gaussian mixture model and allow each attention head to focus on different parts of the input sequence efficiently. Compared to its conventional transformer counterpart, Transformer-MGK accelerates training and inference, has fewer parameters, and requires fewer FLOPs to compute while achieving comparable or better accuracy across tasks. Transformer-MGK can also be easily extended to use with linear attention. We empirically demonstrate the advantage of Transformer-MGK in a range of practical applications, including language modeling and tasks that involve very long sequences. On the Wikitext-103 and Long Range Arena benchmark, Transformer-MGKs with 4 heads attain comparable or better performance to the baseline transformers with 8 heads.en_US
dc.language.isoen_USen_US
dc.titleImproving transformers with probabilistic attention keysen_US
dc.typeArticleen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

  • Le Duy Dung, PhD [3]
    Assistant Professor, Computer Science program, College of Engineering and Computer Science

Show simple item record


Vin University Library
Da Ton, Gia Lam
Vinhomes Oceanpark, Ha Noi, Viet Nam
Phone: +84-2471-089-779 | 1800-8189
Contact: library@vinuni.edu.vn