Rakshit Agrawal, Principal AI Scientist at Synthpop AI, presents the “Transformer Networks: How They Work and Why They Matter” tutorial at the May 2025 Embedded Vision Summit.
Transformer neural networks have revolutionized artificial intelligence by introducing an architecture built around self-attention mechanisms. This has enabled unprecedented advances in understanding sequential data, such as human languages, while also dramatically improving accuracy on nonsequential tasks like object detection.
In this talk, Agrawal explains the technical underpinnings of transformer architectures, from input data tokenization and positional encoding to the self-attention mechanism, which is the core component of these networks. He also explores how transformers have influenced the direction of AI research and industry innovation. Finally, he touches on trends that will likely influence how transformers evolve in the near future.
See here for a PDF of the slides.

