Abhi I.’s Post

View profile for Abhi I., graphic

OpenAI Unveils GPT-4o: The new battle for the gateway UI ! OpenAI demoed the capabilities of their newest flagship model GPT-4o. They are making it available through a new updated UI, and also through the API. GPT-4o ("Omni") - is a conversational model that improves on its capabilities over text, audio and vision. This new model reasons across voice, text and vision instead of just of just combining them. GPT-4o can be combined with: real-time vision (both still images and videos), memory, GPTs, browse, and advanced data analysis. This makes GPT-4o way faster (2x) and crucially cheaper per token (50%) with a 5x faster rate limit compared to GPT4. But the most powerful impact is how conversational and natural it sounds with the ability to use different tones and even continue post an interruption as a result of understanding the context and the tone of the conversation. The demo was eye-opening with no lag. Demos ranged from math tutoring to telling bed-time stories and real time translation from Italian to English. The model was clearly picking up on emotions and able to change tone and voice to demonstrate emotions. What was magical, and just slightly creepy, was it gave off the appearance of enjoying the interaction! Those of you who were fascinated by the movie "Her" - know where this could go! With this demo, OpenAI has put all other providers of natural language assistants on notice - don't be surprised to see a fire back from everyone else in this space - Apple, Microsoft, Google among others. In a bid to make ChatGPT even more ubiquitous, they rolled out a desktop version of ChatGPT, and an improved Web UI. This feels like shades of the last browser battle playing out all over again. If you recall that bygone era - Netscape pioneered that paradigm, only to be caught and passed by Microsoft, Google & Apple all with their own entrants who bundled this with each of their devices and Operating systems. Mira Murati who led the introduction also emphasized their focus on safety, not a surprise given the drama over the last year. In terms of implications for others, what comes to my mind is the interview that Sam Altman and Brad Lightcap gave recently. (I am paraphrasing) "If your company builds on a foundational model and you cheer a 10x improvement, then you likely building the right way and have a sustainable business model, if it sends a chill down your spine - you may want to reformulate your product/company thesis". Would love to hear your thoughts on this announcement - do you agree that this represents a browser like moment to capture the new UI gateway for the masses? Who is most at risk? Who benefits the most? What would you do differently as a result of what you heard on the demo? #genai #openai

Introducing GPT-4o

https://www.youtube.com/

Thanks for sharing, Abhi. This is great for an improved experience. As a general user, the value-add will come primarily in two forms: faster responses during peak hours and more effective engagements. Often, it is necessary to refine queries to get the optimal experience with GPT-4. Hopefully, the experience with the turbo version will be more exciting and engaging, with enhanced empathy and emotional intelligence to better interpret the tone during human interactions. Exciting indeed...

Nasir Mahmood

Generative AI Consultant | Ex-PwC, Deloitte, AWS, Accenture Executive | Trusted Advisor to Fortune 100 Companies.

2mo

Thanks for sharing, Abhi I.. I've been using 4o since its recent release, and the results have been truly astonishing in comparison.

Multimodality is a huge game changer imo, exciting (and scary) times ahead! And hope they bring Scarlet Johansen’s voice back 😁

See more comments

To view or add a comment, sign in

Explore topics