#!/usr/bin/env python
# coding: utf-8

# Relative multi-head attention

# In[ ]:


import torch
from torch import nn
from labml.logger import inspect
from attention_impl import MultiHeadAttention

