Simon Willison · · 1 min read

llm 0.32a2

Mirrored from Simon Willison for archival readability. Support the source by reading on the original site.

12th May 2026

Release llm 0.32a2 — Access large language models from the command-line

A bunch of useful stuff in this LLM alpha, but the most important detail is this one:

Most reasoning-capable OpenAI models now use the /v1/responses endpoint instead of /v1/chat/completions. This enables interleaved reasoning across tool calls for GPT-5 class models. #1435

This means you can now see the summarized reasoning tokens when you run prompts against an OpenAI model, displayed in a different color to standard error. Use the -R or --hide-reasoning flags if you don't want to see that.

Posted 12th May 2026 at 5:45 pm

Recent articles

This is a beat by Simon Willison, posted on 12th May 2026.

projects 528 ai 2018 annotated-release-notes 49 openai 419 generative-ai 1786 llms 1752 llm 599

Monthly briefing

Sponsor me for $10/month and get a curated email digest of the month's most important LLM developments.

Pay me to send you less!

Discussion (0)

Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.

Sign in →

No comments yet. Sign in and be the first to say something.

More from Simon Willison