Blog
5 hours ago
The Simplest Way to Understand How LLMs Actually Work!
Transformers use a clever trick: for every word (technically tokens), the model creates three different representations: Q, K, V. The model compares the query from one word against the keys of all other words. This produces ATTENTION SCORES
Source: HackerNoon →