Splitting off from #14721 > FYI a start has been made here https://github.com/vllm-project/vllm/pull/20504 > > That PR (which was merged to `main` on [7/9/2025](https://github.com/vllm-project/vllm/pull/20504#event-18495144925)) explicitly has an unchecked boxes for > > * [ ] Tool/functional calling support > * [ ] Output token streaming > > Any plans to implement those features? I think that is what is needed to support agentic coding tools like codex. See: > > * https://docs.vllm.ai/projects/recipes/en/latest/OpenAI/GPT-OSS.html#harmony-format-support _Originally posted by @bartlettroscoe in [#14721](https://github.com/vllm-project/vllm/issues/14721#issuecomment-3321963360)_