
AI that truly understands video. Uses multimodal models (Marengo/Pegasus) to search, analyze & generate text from video content at scale.
AI that truly understands video. Uses multimodal models (Marengo/Pegasus) to search, analyze & generate text from video content at scale.
Hi everyone!
Co-hunt with @sentry_co to share TwelveLabs, an AI platform focused entirely on deep video understanding – going beyond just frame analysis or basic metadata.
The name itself looks familiar in the current AI audio/voice landscape... 😉 But it's actually something about video AI, aims to let you interact with video content in powerful new ways.
Powered by their multimodal models (Marengo for search/embeddings, Pegasus for text generation), you can use natural language to semantically search for moments within large video libraries, generate detailed summaries or analyses, and create rich embeddings for tasks like classification or recommendations.
It handles large-scale libraries and offers a free tier and playground to get started. Seems like a robust platform for anyone working seriously with video data.
Raycast
@sentry_co @zaczuo wen ThirteenLabs?
@chrismessina Very soon👀
@chrismessina @zaczuo Audio -> Video -> VR?
Wow man! It's really impressive. You're bringing video understanding to the next level and I'm sure many content creators will love this. Wish you all the best!
honestly feels like Tony Stark would use to analyze CCTV footage. Might mess around and find my lost sunglasses from 2020.