Vivek Gupta Profile picture
Apr 30 10 tweets 4 min read Read on X
#NAACL2025

(1/8) Excited to present our paper "Leveraging LLM For Synchronizing Information Across Multilingual Tables" at #NAACL2025! 🎉

Tackling the challenge of outdated info in low-resource language Wikipedia tables, we propose a novel LLM-based zero-shot approach using task decomposition.💡

🔗 Check out our work here: zero-shot-llm-infosync.github.io/zero-shot-llm-…Image
(2/8) - The problem is huge; the language gap is real.

🌍 300+ languages on Wikipedia

🇬🇧 English dominates: 23% of articles, 76% of edits

🌐 Outside the top 25 languages, active editors, pages, and edits make up <1%.

#Wikipedia #LanguageBarrier #DigitalDivide Image
(3/8) We tackle the problem of information mismatch in tables for different languages, where data can be outdated, unmapped, or missing cultural context.

This is a significant challenge given the vast number of languages on platforms like Wikipedia.
(4/8) 🚀 Proposing a 2-step approach:
🔄 Information Alignment
📈 Information Updation

📝 Prior work (Khincha et al., ACL 2023) missed info updation and didn’t leverage SOTA LLMs for these tasks! #AI #MachineLearning #LLMs #Innovation

We propose a new dataset for updating tables, using past versions as outdated references. #AI #DataScience #TableUpdateImage
(5/8) For information alignment, we use InfoSync (Khinch et. al. 2023 ACL Findings) and combine with LLMs ensemble to enhance performance. Image
(6/8) For information updation - Our paper, explores how state-of-the-art LLMs can enhance the synchronization of multilingual entity-centric tables.

🔄 Hierarchical Task Decomposition Prompt
Break tasks into multiple prompts for sequential inference: Translation, Knowledge graph conversion, Merging, Update from KG, and Back-Translation
#TaskDecomposition

We investigate zero-shot prompts and how to improve LLM performance in this task.Image
(7/8) Key contributions include using SOTA LLMs for information alignment and introducing InfoUpdate dataset, the first human-annotated benchmark for information updation.

We propose a multi-step decomposition zero-shot LLM approach that boosts information factuality and coherence across multilingual entity-centric tables..
(8/8) Catch our oral presentation (I'm presenting) on Wednesday, April 30th, during Session D: Oral/Poster 3, in Ballroom B, at 4 PM.

Join work with Sidharth Khincha, Ankita Anand, Tushar Kataria, @DanRothNLP

@SCAI_ASU, @IITGuwahati , @UUtah , @cogcomp

Looking forward to seeing you there! 👋🗓️📍 #NLP #LLMs #Multilingual #InformationSynchronization
7.1 - we also introduce a new metric for evaluation, based on tri-align, bi-align, un-aligned rows. Image
Image
7.2 We outperform all other traditional baseline, even ones with multiple prompts

We add more information, correct outdated information, and also reduce hallucinated information. Image

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Vivek Gupta

Vivek Gupta Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(