Google's largest and most capable AI model. Built from the ground up to be multimodal, Gemini can generalize and seamlessly understand, operate across and combine different types of information, including text, images, audio, video and code.
This is great but like it goes with all other Google AI products, it needs to be launched in public and tested by devs in real-time to see how it fares against other top LLMs. Ultra is not going to be released soon and as a developer, I'm mostly reluctant to switch to that unless it's significantly better than GPT-4 because of the slower pace of updates and the potential likeliness of Google killing the model in a couple of years.
We use Vertext AI in our sustainable travel planner www.biliki.ai . Looking forward to 13th December when the new Gemini model will be added to Vertex. Let's meet a new era together.
Hi guys, congrats on launch, checked out how Gemini vision works on Bark - there is room for improvement. Uploaded a pic of sofa in living room - Gemini detected sofa and door and tv and lots of other stuff there was not on the pic :)
Nice 🚀💯
Now we use OpenAI API for our AI Furniture Planner. But it will be very interesting try the same in Gemini.
ps. Today is AI Furniture Planner Launch on PH 👋
Gemini, as Google's answer to GPT-4, seems like a significant advancement in AI technology. Wishing the team success! How does Gemini's multimodal capabilities enhance its performance in understanding and operating across various types of information?
Pretty weak honestly and can feel the management pushing lots of weird things out there. Done are times when google could do and push out great things.
Replies
ShortVideoGen
AIWatchfulCompanion
Upcoming
Kommunicate
Scade.pro
Paper Billionaires
SprintsQ
Scade.pro
Akriti! 100+ Abstract Vector Shapes
Huudle AI Project Assistant