This isn't deep learning. But it'll be interesting if you do machine learning, like incomplete information games or play #poker.
π§΅π
2/8) Kuhn Poker is a simple 2-player betting game with three cards (A, K, Q). A single card is dealt to each player. Players take turns betting chips and the player with the higher card wins the chips. If a player folds the other player wins the chips.
π
3/8) CFR finds the Nash equilibrium with self-play. In each iteration, it calculates the regret of following the current strategy instead of playing each action. Then it updates the strategy with regret matching:
strategy = regret of action/total regret of all actions
π
4/8) The average of the strategies throughout the iterations gets close to the Nash equilibrium as we iterate.
Nash equilibrium is a state where no player can increase their expected payoff by changing their strategy.
π
5/8) The strategy is a function of "information set" and gives a probability distribution across actions. An "information set" is the state of the game thatβs visible to the player.
π
6/8) Our implementation is accompanied by a lengthy introduction to CFR and MCCFR. The MCCFR implementation is abstracted from the game Kuhn Poker and we will add Leduc Poker implementation soon.
It identifies research papers mentioned in websites you visit and shows a 2-line summary, availability code/videos/discussions, popularity on Twitter, and conferences.
The paper introduces ConvMixer which mix patch embeddings with depth-wise and point-wise convolutions.
π§΅π
2/ The implementation is very simple and the paper presents a 280 character version of the @PyTorch model code - fits a tweet πͺ
Our implementation is a bit lengthy (hopefully easier to understand π)
π
3/ ConvMixer is similar to MLP-mixer but uses linear transforms (convolutions) instead of multiple layers for each mixing. And It only mixes the neighboring patches within the convolution kernel.
2/ This removes noise (denoise) step-by-step to generate images. It adds noise to an image from the dataset iteratively and a model is trained to predict the noise at each step.