I’m a bit late, but check out @Apple research papers presented at #ICML2022 here: machinelearning.apple.com/research?year=…
Includes work on transformer pre-training, dynamic pooling, private federated statistics, and generative models.
1/n
arxiv.org/abs/2207.07611 provides a simple method for transformer pre-training based on predicting the positions of a set of orderless tokens. Often requires fewer pre-training iterations to achieve a good model.
2/n
Xavier Suau, Luca Zappella, and Nick Apostoloff introduced arxiv.org/abs/2110.02802 for guiding language generation with expert units derived from latent transformer features, helping to study and mitigate bias in generated text.
7/n
Finally, colleagues in AI/ML along with #UCSD authors introduced a new method for controlling style and content for handwritten text in arxiv.org/abs/2110.02891. 8/n
I hope this is helpful for those who may have missed the #ICML2022 presentations or wanted to se a more comprehensive list. Enjoy!
n/n
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Look! Realistic scene generation with a freely moving camera! Compare the effect of constrained vs free viewing. This is an effort we envisioned several years ago when thinking about the power of neural nets to learn expressive priors about the world. 1\4
If we can generate realistic scenes from a prior, we can start to understand what it means to ‘see’ within the context of a scene, by inversion (analysis-by-synthesis). We wanted to do this with realistic scenes to ensure that GSN really understand the 3D world. 2\4
I’m excited to put GSN to use to fuel new applications in model-based RL, AR/VR, and scene understanding. 3/4