Many models bake in domain knowledge to control how input data is processed. This means models must be redesigned to handle new types of data.
Introducing the Perceiver, an architecture that works on many kinds of data - in some cases all at once: dpmd.ai/perceiver (1/)
Like Transformers, Perceivers process inputs using attention. But unlike Transformers, they first map inputs to a small latent space where processing is cheap & doesn’t depend on the input size. This allows us to build deep networks even when using large inputs like images. (2/)
Perceivers can learn a different attention pattern for each type of data (shown for images and video), making it easy for them to adapt to new data and unexplored problems where researchers may not know what kinds of patterns they should be looking for. (3/)
In a new paper, our team tackles a fundamental AI problem: how can we simultaneously parse the world into objects and properties, while simultaneously inducing the rules explaining how objects change over time: dpmd.ai/3fmrxsn (1/)
Work by @LittleBimble with @pfau, @pushmeet, Matko Bosnjak, Lars Buesing, Kevin Ellis, and Marek Sergot. (2/)
This system combines the Apperception Engine with a binary neural network to learn a provably 100% accurate model of non-trivial environments (e.g. Sokoban) from noisy raw pixel data. (3/)
Multimodal transformers achieve impressive results on many tasks like Visual Question Answering and Image Retrieval, but what contributes most to their success? dpmd.ai/3h8u23Z (1/)
This work explores how different architecture variations, pretraining datasets, and losses impact multimodal transformers’ performance on image retrieval: dpmd.ai/3eENAtF
Multimodal transformers outperform simpler dual encoder architectures when the amount of data is held constant. Interestingly, larger datasets don’t always improve performance. (3/)
AI has been extremely successful in real world games (GO, DOTA, StarCraft) with results coming from relatively simple multi-agent algorithms. In this paper, we hypothesise that they share a common geometry - Spinning Tops. Learn more: bit.ly/3qI8RrD#NeurIPS2020
I’ve always loved biology. During my masters I decided to take a handful of neurophysiology courses - which I found to be interesting. But eventually I realised that my true strengths were in mathematical sciences. A career in ML and AI became a natural way to combine the two.
Yesterday we shared the news that #AlphaFold has been recognised as a solution to the ‘protein folding problem’ by #CASP14, the biennial Critical Assessment of Protein Structure Prediction. But what exactly is protein folding, and why is it important? A thread… (1/6)
Proteins are the building blocks of life - they underpin the biological processes in every living thing. If you could unravel a protein you would see that it’s like a string of beads made of a sequence of different chemicals known as amino acids. (2/6)
Interactions between these amino acids make the protein fold, as it finds its shape out of almost limitless possibilities. For decades, scientists have been trying to find a method to reliably determine a protein’s structure just from its sequence of amino acids. (3/6)
He suggests the Deep Learning Book from @mitpress for a comprehensive introduction to the fundamentals of DL: bit.ly/351qMzb (1/7)
Overwhelmed with the number of available machine learning courses? @seb_ruder recommends taking a look through @venturidb’s curated - and ranked - list available on @freeCodeCamp.
We’re back with more #AtHomeWithAI researcher recommendations. Next up is research scientist @csilviavr with suggestions for resources to learn about causal inference! (1/5)
Her first suggestion is “The Book of Why” by @yudapearl & Dana Mackenzie.
According to Silvia, this is best for those looking for an introduction to the topic: bit.ly/30isGej#AtHomeWithAI
Need a more in-depth look at causal inference? Silvia suggests reading through “Causal Inference in Statistics: A Primer” by @yudapearl, @MadelynTheRose & @NP_Jewell.