🔥Excited to release LLaMA-Adapter! With only 1.2M learnable parameters and 52K instruction data, LLaMA-Adapter turns a #LLaMA into an instruction-following model within ONE hour, delivering high-quality responses!
We adopt learnable adaption prompts and prepend them to the input text tokens at higher transformer layers. A zero-init attention mechanism with zero gating adaptively injects the new instructional cues into LLaMA, while effectively preserving its pre-trained knowledge.
With efficient training, LLaMA-Adapter generates high-quality responses, comparable to Alpaca with fully fine-tuned 7B parameters.
LLaMA-Adapter can be simply extended to multi-modal input, e.g., images, for image-conditioned LLaMA, which achieves superior reasoning capacity on #ScienceQA (scienceqa.github.io), a recent multi-modal science question benchmark.
With a mere 1.2 million learnable parameters, LLaMA-Adapter demonstrates superior reasoning capacity on #ScienceQA, surpassing a diverse range of multi-modal and LLM models, such as fully-finetuned MM-COT and few-shot GPT-3.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
🧵1/n Here's an example from ScienceQA: . TextGrad reviews the multimodal question and provides evaluations to “propagate” the initially incorrect answer to the correct one. scienceqa.github.io
🧵2/n TextGrad refines misinterpretations in the final conclusion for this biology question through automatic “differentiation” via text.
🎉Exciting news: LLaMA-Adapter is now fully unlocked! 🧵6
1⃣ As a general-purpose #multimodal foundation model, it integrates various inputs like images, audio, text, video, and 3D point clouds, while providing image, text-based, and detection outputs. It uniquely accepts the… twitter.com/i/web/status/1…
🧵1/6 Experience the magic of LLaMA-Adapter! Transforming real-world inputs like text, images, videos, audio, and 3D point clouds into engaging text. The reality you know, reimagined through AI.
🖼️📽️🔉🌐➕📝 ➡️➡️🦙➡️➡️ 📝
🧵2/6 LLaMA-Adapter goes beyond creating text! It's also capable of generating detection results, bringing a new dimension to understanding and interacting with the world.
🖼➕📝 ➡️➡️🦙➡️➡️ 📝➕🔍
🚨Struggling to select examples for GPT-3? Try our PromptPG, the first work that applies RL to select in-context examples for GPT-3! PromptPG achieves a gain of 5.31% on TabMWP, a new dataset of tabular math word problems! Check out data and codes:👇 promptpg.github.io 🧵1/7