[THREAD] Following the public release of Spaces, here is a showcase of a few ones we like. Let’s start with this surprising Draw-to-Search demo by @osanseviero and powered by CLIP. huggingface.co/spaces/osansev…
What a time to be alive! You can finally decode your doctor's prescription with this cool OCR demo by @NielsRogge - using the Microsoft TrOCR encoder-decoder model. huggingface.co/spaces/nielsr/…
Part 1 of the course focused on text classification, part 2 will focus on all other common NLP tasks. @mervenoyann has made videos to introduce you to each of them!
Let's start with Token Classification (giving a label to some/each word in a sentence):
Then there is question answering: finding the answer to a question in some context.
Next is Causal Language Modeling: guessing the next word in a sentence. This is how GPT-2 and its descendants were pretrained.
Participants of the SUPERB Challenge can submit their results to the AAAI 2022 Workshop: The 2nd Self-supervised Learning for Audio and Speech Processing🤖
The winners will be invited to present their methods 🏅🏅🏅!
Starting today all 🤗 Spaces are publicly viewable 🚀 You can find all the amazing demos created as part of the sprint here 👉 huggingface.co/spaces
This has been the largest Hugging Face event, and we're extremely excited by the results. Almost 800 members joined and had almost 100 projects, 170 models & 36 Spaces! 🤯 That is super impressive given the timeframes of the event!
2. Having read the explanation, if this is a project that interests you and that you think you will be able to finish within ~6 weeks - obviously with the help of the Hugging Face team - please send us a message to team@huggingface.co
Blog alert: check out the new guest post by Amog Kamsetty and the @raydistributed team on training a Retrieval Augmented Generation Model with Hugging Face and Ray!
The RAG model by @olapiktus@PSH_Lewis and @facebookai colleagues leverages external knowledge sources like Wikipedia to have direct and dynamic access to information at inference time
Part of this process relies on training a retriever to learn how to find that information
@raydistributed is a framework-agnostic, flexible implementation for ad-hoc concurrent programming, which makes it ideal for scaling up this training, making retrieval 2x faster and drastically improving the scalability of RAG distributed fine-tuning