by Collin Flynn | at Minnebar 18 | 1:00 โ 1:45 in Minnetonka | View Schedule
"Transformer" is a very generic name for a specific machine learning architecture, first published in a paper titled "Attention Is All You Need".
Transformers were conceived to handle text translation, but are surprisingly effective in other (non-text) domains. What was the key innovation that set transformers apart from previous methods?
In this talk we'll look at embeddings, gradient decent, loss functions, and how Transformers learn to pick up on conversational context using matrix multiplication. We'll also take a look at a newer architecture called "Mamba", and how it improved upon some of weaknesses of Transformers.
Collin is a software developer with Livefront
Does this session sound interesting? You may also like these:
This will add your name to the list of interested participants. It will help us gauge interest for scheduling purposes.