Since its debut in Neural Machine Translation(NMT), attention mechanism has been witnessed widely and frequently in various studies of Natural Language Processing(NLP) and Computer Vision(CV). As I am unfamiliar with CV, here I will only talk a bit about what attention is by intuition and in implementation, how it works and hopefully why it proves useful.
The editor converts LaTeX equations in double-dollars $$
: . All equations are rendered as block equations. If you need inline ones, you can add the prefix
\inline
: . But it is a good practice to place big equations on separate lines: