OpenAI has rolled out server-side changes to GPT-5 that markedly speed up the model’s responses in Cursor, a popular AI-assisted code editor. Developers testing the updated interface report that 95th-percentile latency is now roughly twice as fast after improvements to caching and API throughput. The upgrade arrives alongside a pricing revision that reduces the cost of cached or repetitive input tokens to one-tenth of the standard rate, down from one-quarter previously. Users say the lower fee makes long programming sessions and automated refactoring significantly cheaper. While the base GPT-5 model now executes faster, developers note that capability varies by subscription tier. ChatGPT Plus and Team accounts provide a smaller context window and lower "GPT-5 Thinking" setting than Pro accounts, which continue to offer the higher-reasoning configuration used in Cursor’s “gpt-5-high” option.
On the consilience of evidence to date, I think I've been correct that Grok 4 is the first «next generation model», it's on the same tier as GPT-5, and its shortcomings are mainly because xAI has little experience building and shipping LLM products (and accumulated good data). https://t.co/Z4SJrRzXZD https://t.co/SLeBazFCUO
It appears that base GPT-5 is actually a small/medium size model but it makes up for this downside in thinking provess . GPT-5-thinking high is unrecognizably better than the base model. It also appears Opus 4.1 is the largest thinking model in existence and it is exceptionally
GPT-5 thinking in plus tier is not the same as GTP-5 thinking in pro tier. That „thinking“ (reasoning) setting is higher. https://t.co/UQkHLHhWs2 https://t.co/ZYXEVB8a9T