Papers with Code Profile picture
Aug 31, 2022 12 tweets 6 min read Read on X
🔥Top Trending ML Papers of the Month

Here is a thread to catchup on the top 10 trending papers of August on @paperswithcode. Image
1) An Image is Worth One Word - a new approach that allows for more creative freedom with image generation; proposes "textual inversions" to find pseudo-words that compose new sentences that guide personalized creations.

paperswithcode.com/paper/an-image…
2) Cold Diffusion - proposes diffusion models built around arbitrary image transformations without Gaussian noise; discusses the potential for generalized diffusion models that invert arbitrary processes.

paperswithcode.com/paper/cold-dif… Image
3) Image as a Foreign Language - proposes a multimodal foundation model called BEiT-3 which achieves state-of-the-art performance on many vision and language tasks.

paperswithcode.com/paper/image-as… Image
4) 3D Vision with Transformers - a comprehensive overview of transformers for 3D tasks, which include classification, segmentation, detection, pose estimation, and more.

paperswithcode.com/paper/3d-visio… Image
5) LLM.int8() - a new quantization procedure that allows large scale model checkpoints (16/32-bit) to be loaded and converted to Int8. This allows access to large language models (LLMs) that could not be accessed due to limited GPU memory.

paperswithcode.com/paper/llm-int8… Image
6) Flow-Guided Transformer - proposes a Transformer-base model leveraging motion discrepancy from optical flows; this approach helps to instruct attention retrieval in transformers for video inpainting.

paperswithcode.com/paper/flow-gui… Image
7) MinVIS - a minimal video instance segmentation framework, without video-based training, that produces state-of-the-art performance and is comparable to fully-supervised approaches.

paperswithcode.com/paper/minvis-a… Image
8) PeRFception - leverages NeRF variant to create large-scale implicit representation datasets for perception tasks.

paperswithcode.com/paper/perfcept… Image
9) YOLOPv2 - an effective and efficient multi-task learning network for performing faster and better on tasks such as traffic object detection and lane detection.

paperswithcode.com/paper/yolopv2-… Image
10) Deep Patch Visual Odometry - a new deep learning system for monocular visual odometry that achieves 2x-5x real-time speeds; outperforms previous works on several benchmarks in terms of accuracy and speed.

paperswithcode.com/paper/deep-pat… Image
You can catch more trending machine learning papers daily on Papers with Code. Check out the following sections:

Top Trending: paperswithcode.com
Top Social: paperswithcode.com/top-social
Browse SoTA: paperswithcode.com/sota

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Papers with Code

Papers with Code Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @paperswithcode

Nov 15, 2022
🪐 Introducing Galactica. A large language model for science.

Can summarize academic literature, solve math problems, generate Wiki articles, write scientific code, annotate molecules and proteins, and more.

Explore and get weights: galactica.org
We believe models should be open.

To accelerate science, we open source all models including the 120 billion model with no friction. You can access them here.

github.com/paperswithcode…
We release our initial paper below. We train on a large scientific corpus of papers, reference material, knowledge bases and many other sources. Includes scientific text and also scientific modalities such as proteins, compounds and more.

galactica.org/paper.pdf
Read 7 tweets
Jul 19, 2022
Keeping up with Language Models

Check out these trending papers to catchup on the latest developments in language models. ↓
1) N-Grammer (Roy et al.) - takes inspiration from statistical language modeling and augments Transformers with latent n-grams; it matches strong baseline models like Transformer and Primer while being faster in inference.

paperswithcode.com/paper/n-gramme…
2) Language Models (Mostly) Know What They Know (Kadavath et al.) - investigates whether an LM can be trained to perform well at predicting which questions it will be able to answer correctly; this enables self-evaluation on open-ended sampling tasks.

paperswithcode.com/paper/language…
Read 8 tweets
Jul 5, 2022
🔥Top Trending ML Papers of the Month

Here is a thread to catchup on the top 10 trending papers of June on @paperswithcode. ↓
1️⃣ Mask DINO (Li et al) - extends DINO (DETR with Improved Denoising Anchor Boxes) with a mask prediction branch to support image segmentations tasks (instance, panoptic, and semantic).

paperswithcode.com/paper/mask-din…
2️⃣ Hopular (Schäfl et al) - proposes a deep learning architecture based on continuous Hopfield networks for competitive results on small-sized tabular datasets.

paperswithcode.com/paper/hopular-…
Read 12 tweets
May 31, 2022
🔥Top Trending ML Papers of the Month

Here is a thread to catchup on the top 10 trending papers of May on @paperswithcode. 1/11
1⃣ OPT (Zhang et al) - release open pre-trained transformer language models ranging from 125M to 175B parameters. The release include: logbook detailing infrastructure challenges and code to experiment with the released models. 2/11

paperswithcode.com/paper/opt-open…
2⃣ CoCa (Yu et al) - a new foundation model that achieves new state-of-the-art on ImageNet (90.6%); proposes minimal strategy to jointly pre-train an image-text encoder decoder with contrastive loss and captioning loss. 3/11

paperswithcode.com/paper/coca-con…
Read 12 tweets
Apr 25, 2022
10 Recent Trends in Language Models

In this thread, we summarize ten recent trends and insights in language models. ↓
1) Scaling Laws

Kaplan et al. report that language models (LMs) performance improves smoothly when increasing model size, dataset size, and compute. Recent works provide empirical evidence that LMs are underexplored and can be improved in other ways.

paperswithcode.com/paper/scaling-…
2) Compute-Optimal Models

Hoffmann et al find that large LMs are undertrained and that for a compute-optimal model, Chinchilla, model size & number of training tokens should be scaled equally. Chinchilla (70B) outperforms Gopher (280B) on several tasks.

paperswithcode.com/paper/training…
Read 12 tweets
Apr 12, 2022
Announcing Best Paper Awards for ML Reproducibility Challenge 2021!

We had over 100+ submissions and we are happy to accept 43 reports in our main program. Congratulations to our best and outstanding paper award winners. See more here: paperswithcode.com/rc2021 Image
Our program would not be possible without the support of our awesome reviewers! To honor their hard work, we are excited to announce the Outstanding Reviewer Awards! Image
Stay tuned for more updates regarding the release of ReScience journal, and our plans for a one-day workshop on Reproducibility where we showcase these reports. Image
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(