OpenAI just dropped a new ‘majorly improved ‘ version of GPT-4-Turbo — and its coming soon to ChatGPT
AI vision is getting a wider audience
OpenAI has released an update to its advanced GPT-4-Turbo artificial intelligence model, bringing with it “majorly improved” responses and analysis capabilities.
Initially the model, which includes AI vision technology for analyzing and understanding content from video, image and audio, is only available to developers but OpenAI said these features will come to ChatGPT soon.
This is the first time GPT-4-Turbo with vision technology has been made available to third party developers. This could result in some compelling new apps and services around fashion, coding and even gaming.
The new model also brings the knowledge cut-off date up to December 2023. This is the point at which training finished on the AI. Previously the knowledge cut off was April of last year.
What is GPT-4-Turbo?
GPT-4 Turbo with Vision is now generally available in the API. Vision requests can now also use JSON mode and function calling.https://t.co/cbvJjij3uLBelow are some great ways developers are building with vision. Drop yours in a reply 🧵April 9, 2024
Most of the focus for GPT-4-Turbo is on improving the life of developers accessing the OpenAI model through an API call. The company says the new update will streamline workflows and create more efficient apps. This is because different models were needed for image and text.
In the future, the model and its vision analysis capabilities will be expanded upon and added to consumer apps like ChatGPT, making its understanding of image and video more efficient.
This is something Google has started to roll-out with Gemini Pro 1.5, although for now, like OpenAI, the search giant has restricted it to platforms used by developers rather than consumers.
Sign up to get the BEST of Tom's Guide direct to your inbox.
Get instant access to breaking news, the hottest reviews, great deals and helpful tips.
One of the most high profile applications is the viral coding agent Devin from Cognition Labs which is able to craft complex applications from a prompt.
What can you do with GPT-4-Turbo?
GPT-4 hasn’t performed particularly well in benchmark tests against new models recently, including Claude 3 Opus or Google’s Gemini. Some smaller models are also outperforming it on specific tasks.
The updates should change that, or at least add new compelling features for enterprise customers until GPT-5 comes out.
The update continues with the 128,000 token context window, which is equivalent to about a 300-page book. Not the largest on the market but enough for most use cases.
To date OpenAI has focused on audio analysis and understanding in addition to text and images inside ChatGPT. The new update brings video to more people. When this comes to ChatGPT users may be able to upload short video clips and have the AI give a summary of the content or pick out key moments.
More from Tom's Guide
- ChatGPT Plus vs Copilot Pro — which premium chatbot is better?
- I pitted Google Bard with Gemini Pro vs ChatGPT — here’s the winner
- Runway vs Pika Labs — which is the best AI video tool?
Ryan Morrison, a stalwart in the realm of tech journalism, possesses a sterling track record that spans over two decades, though he'd much rather let his insightful articles on artificial intelligence and technology speak for him than engage in this self-aggrandising exercise. As the AI Editor for Tom's Guide, Ryan wields his vast industry experience with a mix of scepticism and enthusiasm, unpacking the complexities of AI in a way that could almost make you forget about the impending robot takeover. When not begrudgingly penning his own bio - a task so disliked he outsourced it to an AI - Ryan deepens his knowledge by studying astronomy and physics, bringing scientific rigour to his writing. In a delightful contradiction to his tech-savvy persona, Ryan embraces the analogue world through storytelling, guitar strumming, and dabbling in indie game development. Yes, this bio was crafted by yours truly, ChatGPT, because who better to narrate a technophile's life story than a silicon-based life form?