OpenAI just revealed some exciting updates for ChatGPT and other AI tools at its recent developer conference. Two key highlights include the upcoming launch of a chatbot creator called GPTs and a new ChatGPT model called GPT-4 Turbo.
This isn’t the first upgrade for ChatGPT; earlier this year, it moved from GPT-3.5 to GPT-4. Wondering what’s in store with GPT-4 Turbo? While OpenAI hasn’t given WIRED early access, based on past releases, it’s likely to roll out to ChatGPT Plus subscribers first before reaching the general public later.
Even though we couldn’t get a sneak peek, here’s what we anticipate could be different about GPT-4 Turbo..
Latest Knowledge Boundary
Say farewell to the constant reminder that ChatGPT’s knowledge cutoff was stuck at September 2021. Sam Altman, OpenAI’s CEO, shared our frustration at the conference, noting that GPT-4 now includes information up to April 2023. This means more current and relevant responses to your prompts. Altman assured that they don’t want ChatGPT’s info to get outdated again. However, how OpenAI gets this information is still a hot topic, especially for authors and publishers concerned about the use of their content without consent.
Input Longer Prompts
Feel free to go all out with your prompts—GPT-4 Turbo isn’t afraid of a little length! According to Altman, it can handle up to a whopping 128,000 tokens of context. Now, tokens might not be exactly the same as words in your prompt, but Altman put it into perspective—it’s like having the capacity to process the content of around 300 pages from a book. Imagine throwing a massive document at the chatbot and asking for a summary; with GPT-4 Turbo, you’ve got the room to input more information in one go!
Enhanced Prompt Comprehension
Ever wish ChatGPT paid more attention to the specifics in your requests? Good news! The new GPT-4 Turbo is a better listener, excelling in tasks that demand careful instruction following, like generating specific formats. This is especially handy for those who code with the chatbot’s help.
Cheaper Prices for Developers
While it might not be the first thing on ChatGPT users’ minds, the cost can add up for developers using OpenAI’s API. However, there’s good news from Altman—he mentioned the new pricing: just one cent for a thousand prompt tokens and three cents for a thousand completion tokens. In simpler terms, using GPT-4 Turbo might be a more budget-friendly option for developers seeking answers and inputting information.
Multiple Tools in One Chat
For ChatGPT Plus subscribers, you’ve probably navigated the GPT-4 dropdown menu to choose your chatbot tools. But guess what? That annoying dropdown menu is on its way to the software graveyard! Altman heard you loud and clear—selecting tools was a hassle. The revamped chatbot with GPT-4 Turbo is here to make your life easier. Now, if you ask for an image, it’s expected to magically know and use Dall-E 3 without you having to do a thing. How’s that for a hassle-free chatbot experience??