OpenAI has introduced a new beta feature called Flex processing for its API, aimed at providing a more cost-effective option for AI tasks that are less time-sensitive. This new mode is available to developers on tiers 1 through 3 who verify their organizations. Flex processing offers responses at roughly half the price of the standard o3 and o4-mini models but comes with slower response times and occasional resource unavailability. The feature is designed to appeal to users who prioritize budget savings over speed, allowing OpenAI to remain competitive against other AI providers such as Google. Flex processing enables cheaper, slower AI task execution, providing a flexible pricing strategy while maintaining customer engagement.
OpenAI launches Flex processing in beta, an API option halving o3 and o4-mini prices in exchange for slower responses and "occasional resource unavailability" (@kyle_l_wiggers / TechCrunch) https://t.co/AvIOEcIzhJ https://t.co/T0hR6jCumd https://t.co/ZOzeer1FAj
OPENAI’S NEW “FLEX” MODE: CHEAP, SLOW, AND KINDA LAZY OpenAI just launched a budget option for its AI called Flex, and it’s basically the “yeah I’ll do it… eventually” version of their normal models. It costs half as much, but there’s a catch: it’s slower, might ghost you https://t.co/qNDyuFVdya https://t.co/xk3bMQAqKY
This is super interesting news. Kinda creative strategy to make the API pricing more flexible while keeping its customers sticky and staying competitive in AI model landscape. In a bid to more aggressively compete with rival AI companies like Google, OpenAI is launching Flex https://t.co/iP0CD2XPb8