At AWS re:Invent, Twelve Labs announced general availability of Marengo 3.0, a breakthrough video-native foundation model that doesn’t just watch video — it reads dialogue, tracks gestures across minutes, understands emotion, movement, and events through time, and finally making the 90% of digitized data that is video truly searchable and actionable.
Now live on Amazon Bedrock (first cloud provider) and directly via Twelve Labs API.
Unlike stitched-together image+audio models or frame-by-frame analysis, Marengo 3.0 treats video as a single, dynamic system:
Jae Lee, CEO & Co-founder of Twelve Labs: “Video is 90% of digitized data but has been largely unusable. Marengo 3.0 shatters that limit — it’s an incomparable solution for enterprises and developers.”
Nishant Mehta, VP of AI Infrastructure at AWS: “We are excited to be the first cloud provider to offer Marengo 3.0 through Amazon Bedrock, following strong adoption of Twelve Labs’ previous models.”
Start building with Marengo 3.0 → https://twelvelabs.io/marengo
About Twelve Labs
Twelve Labs builds the world’s most powerful video intelligence platform, enabling machines to see, hear, and reason about video like humans. From semantic search to multimodal embeddings, Twelve Labs unlocks the full value of video across media, advertising, government, security, and automotive.