Multi-headed Self-attention Block Implementation
Implementation of a multi-headed self-attention block using only basic PyTorch for use in a vision transformer trained on CIFAR-10
Implementation of a multi-headed self-attention block using only basic PyTorch for use in a vision transformer trained on CIFAR-10