Short Course Q&A Attention in Transformers: Concepts and Code in Py
| Topic | Replies | Views | Activity | |
|---|---|---|---|---|
|
Course presentations
|
|
1 | 7 | December 10, 2025 |
|
Course would not mark the lesson with jupyterlab as completed course Attention in Transformers: Concepts and Code in PyTorch
|
|
3 | 42 | August 29, 2025 |
|
Why do we use square root of key dimension for scaling?
|
|
4 | 128 | July 12, 2025 |
|
The Matrix Math for self-attention
|
|
4 | 106 | February 22, 2025 |
|
In class MaskedSelfAttention -- don't understand python statement
|
|
6 | 107 | February 16, 2025 |