Simon Willison Blog·Frameworks·2d ago·~1 min read
llm 0.32a2
12th May 2026
A bunch of useful stuff in this LLM alpha, but the most important detail is this one:
Most reasoning-capable OpenAI models now use the
/v1/responses
endpoint instead of/v1/chat/completions
. This enables interleaved reasoning across tool calls for GPT-5 class models. #1435
This means you can now see the summarized reasoning tokens when you run prompts against an OpenAI model, displayed in a different color to standard error. Use the -R
or --hide-reasoning
flags if you don't want to see that.
Recent articles
- Notes on the xAI/Anthropic data center deal - 7th May 2026
- Live blog: Code w/ Claude 2026 - 6th May 2026
- Vibe coding and agentic engineering are getting closer than I'd like - 6th May 2026
#fine-tuning#observability
read full article on Simon Willison Blog →