
The most powerful platform for building AI products. Build and scale AI experiences powered by industry-leading models and tools.
The most powerful platform for building AI products. Build and scale AI experiences powered by industry-leading models and tools.
Launched on November 7th, 2023
According to the Verge, Open AI is trying to create a new social network where people can share their creations produced by artificial intelligence.
Meta is also considering creating AI avatars for social media to drive engagement.
On their livestream today, OpenAI just released a bunch of new tools for reliably building and using AI agents. From what I can tell, this is what's new-
New APIs:
Responses API - a new multi-modal API that builds on chat completions to allow for the next-generation of tool calling, starting with the new tools announced today.
I don't see a lot of products using the realtime api in building their conversation ai agents. Given that it now has realtime communication support through WebRTC allowing low latency conversations, I expected it to blow up. Are there any limitations of this model like hallucinations and or is it just too expensive for commercial use?
An incredibly useful tool! Not only does it boost work efficiency, but it also unlocks tons of creative possibilities. I used the following prompt to create my new profile picture, haha! "Transform the character in the attached photo into a 3D chibi-style clay figurine, placed within a Polaroid frame. The visual effect should show the character breaking through the Polaroid's borders, extending into the real world's 2D space. Specifications: Aspect ratio: 9:16 (vertical composition) Character design: Convert the subject into a chibi 3D figurine at 1/4 scale, full-body, retaining original clothing and style with soft, clay-like textures. Background: Extend the Polaroid's background in chibi style (no additional characters), matching the original scene. Polaroid text: Handwritten font: 'XXXX' at the bottom"
Been using ChatGPT lately and notice it's not what it used to be. It keeps missing the point and I find myself explaining things over and over. The answers feel kinda generic now, like it's taking the safe route instead of really getting what I need. Not bad for simple stuff but I expected more progress by now.
Hi everyone!
SWE-Lancer, from OpenAI, is a fascinating new benchmark for evaluating AI models on real-world software engineering tasks. And it's not just about coding – SWE-Lancer also tests AI's ability to make managerial decisions.
This isn't just another synthetic benchmark – it's based on over 1,400 actual freelance jobs posted on Upwork, with a total value of over $1 million.
💰 Real-World Tasks: Everything from small bug fixes to large feature implementations, with associated payouts.
🧑💻 Two Task Types: Coding & Managerial.
🐳 Dockerized: Comes with a unified Docker image for easy setup and consistent evaluation.
🔓 Open-Source: The benchmark data (SWE-Lancer Diamond), Docker image, and evaluation scripts are all open-source.
The idea is to map AI model performance to real-world economic value, for both coding and project management skills. OpenAI's testing shows that even frontier models struggle with many of these tasks.
So, how far are we from the real AI Agent Era?