Google DeepMind Profile picture
Sep 9, 2021 10 tweets 9 min read Read on X
Introducing the '21 DeepMind x @ai_ucl Reinforcement Learning Lecture Series, a comprehensive introduction to modern RL.

Follow along with our researchers are they explore Markov Decision Processes, sample-based learning algorithms & much more: dpmd.ai/2021RLseries 1/2 Image
Also find the full series via the DeepMind @YouTube channel: dpmd.ai/DeepMindxUCL21
In the first lecture of the series, Research Scientist Hado introduces the course and explores the fascinating connection between reinforcement learning and artificial intelligence: dpmd.ai/RLseries1

#DeepMindxUCL @ai_ucl Image
In lecture two, Research Scientist Hado explains why it's important for learning agents to balance exploring and exploiting acquired knowledge at the same time: dpmd.ai/RLseries2

#DeepMindxUCL @ai_ucl Image
In the third lecture, Research Scientist Diana shows us how to solve MDPs with dynamic programming to extract accurate predictions and good control policies: dpmd.ai/RLseries3

#DeepMindxUCL @ai_ucl Image
In lecture four, Diana covers dynamic programming algorithms as contraction mappings, looking at when and how they converge to the right solutions: dpmd.ai/RLseries4

#DeepMindxUCL @ai_ucl Image
In this lecture, Hado explores model-free prediction and its relation to Monte Carlo and temporal difference algorithms: dpmd.ai/RLseries5

#DeepMindxUCL @ai_ucl Image
In part two of the model-free lecture, Hado explains how to use prediction algorithms for policy improvement, leading to algorithms - like Q-learning - that can learn good behaviour policies from sampled experience: dpmd.ai/RLseries6

#DeepMindxUCL @ai_ucl Image
In this lecture, Hado explains how to combine deep learning with reinforcement learning for deep reinforcement learning. He looks at the properties and difficulties that arise when combining function approximation with RL algorithms: dpmd.ai/RLseries7

#DeepMindxUCL @ai_ucl Image
In this lecture, Research Engineer Matteo explains how to learn and use models, including algorithms like Dyna and Monte-Carlo tree search (MCTS): dpmd.ai/RLseries8

#DeepMindxUCL @ai_ucl Image

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Google DeepMind

Google DeepMind Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @GoogleDeepMind

Jun 17
We're sharing progress on our video-to-audio (V2A) generative technology. 🎥

It can add sound to silent clips that match the acoustics of the scene, accompany on-screen action, and more.

Here are 4 examples - turn your sound on. 🧵🔊 dpmd.ai/v2a
✍️ Prompt for audio: “Wolf howling at the moon.”
✍️ Prompt for audio: “A slow mellow harmonica plays as the sun goes down on the prairie.”
Read 5 tweets
May 21
Our video generation model Veo gives more control over the camera. 📹

You can prompt for:
🔘 Extreme close up
🔘 Slow-motion crane shots
🔘 Timelapses

And more. 🧵

✍️ Prompt: “Timelapse of the northern lights dancing across the Arctic sky, stars twinkling, snow-covered landscape.”
✍️ Prompt: “A panning shot of a waterfall cascading down a rocky cliff, lush greenery surrounding the falls, mist rising from the crashing water.”
✍️ Prompt: “A fast-tracking shot down an suburban residential street lined with trees. Daytime with a clear blue sky. Saturated colors, high contrast.”
Read 6 tweets
May 14
Introducing Veo: our most capable generative video model. 🎥

It can create high-quality, 1080p clips that can go beyond 60 seconds.

From photorealism to surrealism and animation, it can tackle a range of cinematic styles. 🧵 #GoogleIO
✍️ Prompt: “Many spotted jellyfish pulsating under water. Their bodies are transparent and glowing in deep ocean.”
✍️ Prompt: “Timelapse of a water lily opening, dark background.”
Read 10 tweets
May 8
Announcing AlphaFold 3: our state-of-the-art AI model for predicting the structure and interactions of all life’s molecules. 🧬

Here’s how we built it with @IsomorphicLabs and what it means for biology. 🧵 dpmd.ai/3URDiNo
AlphaFold 3 can generate the 3D structures of proteins, DNA, RNA, and smaller molecules, while also revealing how they fit together. 🧩

It can also model chemical changes to them that control the healthy functioning of cells - and when disrupted, could lead to disease.
We have also launched AlphaFold Server, a free platform that scientists around the world can use for non-commercial research. 🔬

They can harness AlphaFold 3’s predictions and test hypotheses with just a few clicks - no matter their technical expertise. → alphafoldserver.com
Read 6 tweets
Mar 19
We're announcing TacticAI: an AI assistant capable of offering insights to football experts on corner kicks. ⚽

Developed with @LFC, it can help teams sample alternative player setups to evaluate possible outcomes, and achieves state-of-the-art results. 🧵 dpmd.ai/49PGq1b
📊 Corner kicks can be challenging for AI to model due to the limited availability of data - @premierleague matches only average about 10 a game.

TacticAI uses a geometric deep learning approach to tackle this problem. → dpmd.ai/43p5Gcc
🔍 Analysts need to rewatch many game replays to study rival teams and design future tactics.

TacticAI can help by automatically computing numerical representations of players, allowing them to efficiently look up relevant past routines. ↓ dpmd.ai/49PGq1b
Read 6 tweets
Feb 15
Introducing Gemini 1.5: our next-generation model with dramatically enhanced performance. It also achieves a breakthrough in long-context understanding.

The first release is 1.5 Pro, capable of processing up to 1 million tokens of information. 🧵 dpmd.ai/3SEbw4p
Gemini 1.5 was designed using a new Mixture–of-Experts (MoE) architecture, making it much more efficient to train and serve.

When tested on a set of text, code, image, audio and video evaluations, 1.5 Pro outperforms 1.0 Pro on 87% of benchmarks used for developing our LLMs.
Through a series of machine learning innovations, Gemini 1.5 Pro now has the longest context window of any large-scale foundation model yet.

The bigger the context window, the more information it can take in from a prompt — making its output more consistent, relevant and useful. Static image with text saying: "Introducing Gemini 1.5 Pro."
Read 9 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(