🎥 Building multimodal foundation models for video understanding: https://t.co/iKo4kmNv6p
👉 Join our community: https://t.co/1b53LYUCBc
Oct 24, 2023 • 6 tweets • 2 min read
🚀We are thrilled to release Pegasus-1, our latest video-language foundation model.
We are also excited to share that Twelve Labs has received strategic investments from NVentures of @nvidia, @intelcapital, and @SamsungNext.
Check thread for model details and early access👇
@nvidia @intelcapital @SamsungNext 🎥 We adopt a “Video First” strategy, focusing our model, data, and systems solely on processing and understanding video data.
Our 4 core principles: Efficient Long-form Video Processing, Multimodal Understanding, Video-native Embeddings, Deep Alignment of Video and Language.
Oct 18, 2023 • 17 tweets • 9 min read
🚀What an incredible weekend it was at the @TEDAI2023 23Labs Hackathon!
We are still buzzing from the energy, creativity, and innovation that we witnessed during this two-day event at @SHACK15sf.
@TEDAI2023 @SHACK15sf The enthusiasm for Multimodal AI was immense, with a multitude of projects leveraging text, image, audio, and video modalities to create groundbreaking applications.
It was truly inspiring to see how the participants pushed the boundaries of what's possible with AI.😯