-
Dr Sofia Vallecorsa (CERN)01/02/2024, 09:00
-
Dr Gail Weiss (EPFL)01/02/2024, 11:00
Transformers - the purely attention based NN architecture - have emerged as a powerful tool in sequence processing. But how does a transformer think? When we discuss the computational power of RNNs, or consider a problem that they have solved, it is easy for us to think in terms of automata and their variants (such as counter machines and pushdown automata). But when it comes to transformers,...
Go to contribution page
Choose timezone
Your profile timezone: