We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent c2ffa36 commit 8e2d0e6Copy full SHA for 8e2d0e6
README.md
@@ -1,7 +1,7 @@
1
# Language Modelling Exercise
2
3
This exercsie will allow you to explore language modelling. We focus on the key concept of multi-head attention.
4
-Navigate to the `src/attention_model.py`-file and implement multi-head attention
+Navigate to the `src/attention_model.py`-file and implement multi-head attention [1]
5
6
``` math
7
\text{Attention}(\mathbf{Q},\mathbf{K},\mathbf{V}) = \text{softmax}(\frac{\mathbf{Q}\mathbf{K}^T}{\sqrt{d_k}})\mathbf{V}
0 commit comments