“Transformer Networks: How They Work and Why They Matter,” a Presentation from Synthpop AI

Rakshit Agrawal, Principal AI Scientist at Synthpop AI, presents the “Transformer Networks: How They Work and Why They Matter” tutorial at the May 2025 Embedded Vision Summit.

Transformer neural networks have revolutionized artificial intelligence by introducing an architecture built around self-attention mechanisms. This has enabled unprecedented advances in understanding sequential data, such as human languages, while also dramatically improving accuracy on nonsequential tasks like object detection.

In this talk, Agrawal explains the technical underpinnings of transformer architectures, from input data tokenization and positional encoding to the self-attention mechanism, which is the core component of these networks. He also explores how transformers have influenced the direction of AI research and industry innovation. Finally, he touches on trends that will likely influence how transformers evolve in the near future.

See here for a PDF of the slides.

Here you’ll find a wealth of practical technical insights and expert advice to help you bring AI and visual intelligence into your products without flying blind.

Contact

Address

Berkeley Design Technology, Inc.
PO Box #4446
Walnut Creek, CA 94596

Phone
Phone: +1 (925) 954-1411
Scroll to Top