Multi-headed Self-attention Block Implementation

Implementation of a multi-headed self-attention block using only basic PyTorch for use in a vision transformer trained on CIFAR-10