Batch Transformer Architecture: Case of Synthetic Image Generation for Emotion Expression Facial Recognition
arXiv:2511.11754v1 Announce Type: new Abstract: A novel Transformer variation architecture is proposed in the implicit sparse style. Unlike "traditional" Transformers, instead of attention to sequential or batch entities in their entirety of whole dimensionality, in the proposed Batch Transformers, attention to the "important" dimensions (primary components) is implemented. In such a way, the "important" dimensions or feature selection allows for a significant reduction of the bottleneck size in the encoder-decoder ANN architectures. The proposed architecture is tested on the synthetic image generation for the face recognition task in the case of the makeup and occlusion data set, allowing for increased variability of the limited original data set.
Score: 2.80
Engagement proxy: 0
Canonical link: https://arxiv.org/abs/2511.11754