Reading the Pseudo-code for Transformers
April 10, 2024
![Reading the Pseudo-code for Transformers](https://w1.mtsu.edu/datascience/blog/images/Screenshot_2024_07_15_at_5_02_13___PM_0.png)
Transformers are the underlying algorithms powering large language models like GPT, Bert, and Llama. In this seminar, Dr. Yixiang Wu, with MTSU's Department of Mathematical Sciences will delve into a paper authored by the DeepMind team, focusing on the formal algorithms that define transformers. To grasp these algorithms, attendees need only a basic understanding of linear algebra (matrix multiplications) and probability theory (conditional probabilities).
See the video here.