Compositional de-attention networks

Y Tay, AT Luu, A Zhang, S Wang… - Advances in Neural …, 2019 - proceedings.neurips.cc
… This paper proposes a new quasi-attention that is compositional in nature, ie, learning whether
… Our proposed\textit {Compositional De-Attention}(CoDA) is fundamentally built upon the …

Context-guided bert for targeted aspect-based sentiment analysis

Z Wu, DC Ong - Proceedings of the AAAI conference on artificial …, 2021 - ojs.aaai.org
… Next, we propose an improved Quasi-Attention CG-BERT model that learns a compositional
attention that supports subtractive attention. We train both models with pretrained BERT on …

Structure-aware flow generation for human body reshaping

J Ren, Y Yao, B Lei, M Cui… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
… We employ a compositional attention [22] mechanism for this purpose. Self-attention [26,35]
is initially designed for capturing long-range dependencies. However, it learns to allocate …

An attentive survey of attention models

S Chaudhari, V Mithal, G Polatkan… - ACM Transactions on …, 2021 - dl.acm.org
Attention Model has now become an important concept in neural networks that has been … We
also describe how attention has been used to improve the interpretability of neural networks

Aligning dual disentangled user representations from ratings and textual content

NT Tran, HW Lauw - Proceedings of the 28th ACM SIGKDD Conference …, 2022 - dl.acm.org
… -networks, we employ compositional de-attention mechanism [… networks via a mutual
information maximization strategy. We also propose to leverage compositional deattention network

Fine-Grained Pose Temporal Memory Module for Video Pose Estimation and Tracking

W Chaoyi, Y Hua, T Song, Z Xue, R Ma… - ICASSP 2021-2021 …, 2021 - ieeexplore.ieee.org
… We also transfer compositional de-attention (CoDA) to solve the unique keypoint occlusion
problem in this task and propose a novel keypoint feature replacement to recover the extreme …

Empowering the fact-checkers! automatic identification of claim spans on twitter

M Sundriyal, A Kulkarni, V Pulastya, MS Akhtar… - arXiv preprint arXiv …, 2022 - arxiv.org
… 2019) as our backbone network as it is the best-performing baseline (see Table 4). … All such
compositionally manipulated vectors ZC ij , after interacting with each jth description vectors …

Multi-Representation Variational Autoencoder via Iterative Latent Attention and Implicit Differentiation

NT Tran, HW Lauw - Proceedings of the 32nd ACM International …, 2023 - dl.acm.org
Compositional De-Attention is generally better for refining prototypes than scaled cosine
attention, which supports our hypothesis that CoDA has more representation learning capacity …

Positional Mask Attention for Video Sequence Modeling

J Wang, C Wang, Y Hua, T Song, Z Xue… - … Congress on Image …, 2021 - ieeexplore.ieee.org
… Inspired by the format of the Compositional De-Attention [14], which imports the dissimilarity
representation as to the extra information, we adopt a similar expression to apply the …

Disentangled Graph Variational Auto-Encoder for Multimodal Recommendation With Interpretability

X Zhou, C Miao - IEEE Transactions on Multimedia, 2024 - ieeexplore.ieee.org
… [32], we use a Compositional De-Attention to fuse the information from the two sets. Given
the sets of disentangled representations, the Compositional DeAttention computes an attentive …