We've been on a multi-year effort to take steps towards understanding how well NLP/language tech serves people on a *global* scale. Here's a first report: arxiv.org/abs/2110.06733
We perform meta-analysis of performance across 7 tasks, and devise "global utility" metrics. 1/7
The idea is that language tech should serve every person in the world, not just English native speakers. Based on this, we come up with metrics for language-weighted and population-weighted performance that explicitly consider how many people or languages may benefit 2/7
We then collect performance metrics for seven different tasks, and calculate how well these tasks are doing to serve every language or every population. See some of the breakdowns in the attached figure. 3/7
This allows us to approximate how well a technology is serving potential users throughout the world. It also allows us to identify "pain points," languages that seem to be most underserved, based on our priorities with respect to equity of language or population coverage. 4/7
We also discuss some potential reasons behind current inequities, such as the economic or academic incentives that may cause technology for a particular language to be more or less researched. 5/7
This is a tremendously difficult problem, and the current paper just scratched the surface (with many simplifying assumptions). Nonetheless we (@blasi_lang, @anas_ant, and me) hope this can start a dialog and focus attention/effort on improving technologies globally. 6/7
The overall project has just started and we would definitely love feedback and/or contributions from the broader community! 7/7
• • •
Missing some Tweet in this thread? You can try to
force a refresh
We're building the OpenHands Software Agent SDK, a highly-accurate, flexible, LLM-agnostic solution for building agents for any task you use software for.
1. Coding 2. Maintenance 3. Data processing 4. etc.
We're building the OpenHands Large Codebase SDK, which is specifically aimed at enterprise-ready, large-scale refactoring including app modernization, fixing security issues, improving test coverage, and more.
How far are we from having competent AI co-workers that can perform tasks as varied as software development, project management, administration, and data science?
In our new paper, we introduce TheAgentCompany, a benchmark for AI agents on consequential real-world tasks.
Why is this benchmark important?
Right now it is unclear how effective AI is at helping with real-world work. We hear extreme statements like:
> AI is overhyped, minimally helpful, and doesn’t generalize to new tasks
> AGI will automate all human work in the next few years
This question is hard to answer, but it has implications for:
- Companies: to understand where to incorporate AI in workflows
- Workers: to get a grounded sense of what AI can and cannot do
- Policymakers: to understand effects of AI on the labor market
We compared accuracy across 6 different varieties of tasks:
* Knowledge-based QA (MMLU)
* Reasoning (BIG-Bench Hard)
* Math (GSM8k, SVAMP, ASDIV, MAWPS)
* Code Gen (HumanEval, ODEX)
* Translation (FLORES)
* Web Instruction Following (WebArena)
We tried to control for all variables, using the same prompts, generation params, and evals for all models for fairness. We used:
* @LiteLLM to query models in a uniform way
* @try_zeno to do comprehensive in-depth analysis
All code/data available here: github.com/neulab/gemini-…
CMU Advanced NLP is done for 2022! Check the videos on YouTube 😃
I also rehauled our assignments to reflect important skills in NLP for 2022: github.com/neubig/nlp-fro…
If you're teaching/learning NLP see the 🧵 and doc for more!
Basically, there have been *huge* changes in NLP due to advances BERT and GPT-3. And the skills needed to be a good NLP researcher or engineer have changed too! I've re-designed our assignments to reflect this.
Assignment 1 is now "Build your own BERT", which is a more traditional implementation assignment, building implementation skills and understanding of transformers and the pre-train and fine-tune paradigm.
Retrieval-based models are increasingly important in NLP/QA. But an important factor in modeling text is knowing *where* it came from. Our #ICLR2022 paper proposes retrieval-based LMs considers the "structural locality" of texts to improve retrieval: arxiv.org/abs/2110.02870 🧵↓
We demonstrate this on two example datasets: Wikipedia articles and Java code. We leveraging the article and project structure respectively to define different "locality" levels between two documents.
Our analysis shows that the distance between embeddings, used widely in retrieval tasks, is *not* capturing this locality directly, so further improvements are needed. We do this by learning a function to adjust the distance metric for each locality level in KNN language models.