OpenAI announced it will launch a new AI model, “GPT-4o mini,” the artificial intelligence startup’s latest effort to expand use of its popular chatbot, on Thursday, CNBC reported.
The company called the new release “the most capable and cost efficient small model available today,” and it plans to integrate image, video, and audio into it later.
The mini AI model is an offshoot of GPT-4o, OpenAI’s fastest and most powerful model, which it launched in May during a live-streamed event with executives. The “o” in GPT-4o stands for omni, and GPT-4o has improved audio, video and text capabilities, with the ability to handle 50 different languages at improved speed and quality, according to the company.
OpenAI posted: GPT-4o mini: advancing cost-efficient intelligence
OpenAI is committed to making intelligence as broadly accessible as possible. Today, we’re announcing GPT-4o mini, our most cost-efficient small model. We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable.
GPT-4o mini scores 82% on MMLU and currently outperforms GPT-4(1) on chat preferences in LMSYS leaderboard. It is priced at 15 cents per million input tokens and 60 cents per million output tokens, an order of magnitude more affordable than previous frontier models and more than 60% cheaper than GPT-3.5 Turbo.
GPT-4o mini enables a broad range of tasks with its los cost and latency, such as applications that chain or parallelize multiple model calls (e.g., calling multiple APIs), pass a large volume of context to the model (e.g., full code base or conversation history), or interact with customers through fast, real-time text responses (e.g., customer support chatbots).
Today, GPT-4o mini supports text and vision in the API, with support for text, image, video and audio inputs and outputs coming in the future. The model has a context window of 128K tokens, supports up to 16K output tokens per request, and has knowledge up to October 2023. Thanks to the improved tokenizer shared with GPT-41 , handling non-English text is now even more cost effective…
TechCrunch reported OpenAI introduced GPT-4o mini on Thursday, its latest small AI model. The company says GPT-4o mini, which is cheaper and faster than OpenAI’s current cutting edge AI models, is being released for developers, as well as through the ChatGPT web and mobile app for consumers starting today. Enterprise users will gain access next week.
The company said GPT-4o mini outperforms industry leading small AI models on reasoning tasks involving text and vision. As small AI models improve, they are becoming more popular for developers due to their speed and cost efficiencies compared to larger models such as GPT-4 Omni or Claude 3.5 Sonnet. They’re a useful option for high-volume simple tasks that developers might repeatedly call on an AI model to preform.
In my opinion, it might be better to use a smaller version of GPT-4o than to make larger ones that require things that humans need. More specifically, I have concerns about how much water gets used by AI systems.