Attention mechanisms differ based on where the particular attention mechanism or model finds its application. Another distinction is the areas or relevant parts of the input sequence where the model focuses and places its attention.
The following are the types of attention mechanisms:
Let’s take a closer look at these types.
When a sequence of words or an image is fed to a generalized attention model, it verifies each element of the input sequence and compares it against the output sequence.
So, each iteration involves the mechanism's encoder capturing the input sequence and comparing it with each element of the decoder's sequence.
From the comparison scores, the mechanism then selects the words or parts of the image that it needs to pay attention to.
The self-attention mechanism is also sometimes referred to as the intra-attention mechanism.
It is so-called because it picks up particular parts at different positions in the input sequence and over time it computes an initial composition of the output sequence.
It does not take into consideration the output sequence as there is no manual data entry procedure where the prediction of the output sequence is assisted in any way.
In our next post (part 2), we'll take this discussion forward, discussing the different types of attention mechanisms. So stay tuned!
AI already creates software, hardware is next. Several companies like Circuitmind, Cells and JITX are starting to use AI to do hardware design.
Voxels vs Polygons I like this paper on generating 3D voxels based objects: https://alexzhou907.github.io/pvd As compared to polygon based models, I think voxels are a more accurate way of modeling actual 3D objects. Also seems closer to how 3D printing would work.