Skip to content

Conversation

@aarnphm
Copy link
Collaborator

@aarnphm aarnphm commented Aug 13, 2025

Continuation of #17420

This PR introduces the args --structured-output-config as a way to unify all related structured outputs config in one CLI field.
This would help simplify general UX for specifying custom options with backends.

I also remove all previous guided_decoding options

This would also be considered breaking. There will be no --guided-decoding-* option anymore. Instead, you should use --structured-outputs-config '{...}' or --structured-outputs-config.backend outlines

Signed-off-by: Aaron Pham [email protected]
Signed-off-by: Harry Mellor [email protected]
Co-authored-by: Nick Hill [email protected]
Co-authored-by: Harry Mellor [email protected]

@github-actions
Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

@mergify mergify bot added documentation Improvements or additions to documentation frontend performance Performance-related issues structured-output v1 labels Aug 13, 2025
@mergify
Copy link

mergify bot commented Aug 13, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @aarnphm.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@aarnphm aarnphm force-pushed the feat/decoding-args-rename-all branch from d3ac885 to 69068cd Compare August 13, 2025 00:47
weijinqian0 pushed a commit to weijinqian0/vllm-ascend that referenced this pull request Sep 22, 2025
…llm-project#2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>

- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
mansiag05 pushed a commit to mansiag05/pytorch that referenced this pull request Sep 22, 2025
They have been removed in vllm-project/vllm#25117 and vllm-project/vllm#22772, thus failing in trunk at the moment after the latest pin commit update

Pull Request resolved: pytorch#163383
Approved by: https://github.com/wdvr, https://github.com/seemethere, https://github.com/malfet
Mercykid-bash pushed a commit to Mercykid-bash/vllm-ascend that referenced this pull request Sep 22, 2025
…llm-project#2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>

- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
Signed-off-by: Che Ruan <[email protected]>
Mercykid-bash pushed a commit to Mercykid-bash/vllm-ascend that referenced this pull request Sep 22, 2025
…llm-project#2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>

- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
Signed-off-by: Che Ruan <[email protected]>
cleonard530 pushed a commit to cleonard530/pytorch that referenced this pull request Sep 22, 2025
They have been removed in vllm-project/vllm#25117 and vllm-project/vllm#22772, thus failing in trunk at the moment after the latest pin commit update

Pull Request resolved: pytorch#163383
Approved by: https://github.com/wdvr, https://github.com/seemethere, https://github.com/malfet
@simon-mo
Copy link
Collaborator

@aarnphm can we add backward compatibility for one version so people know how to migrate?

@hmellor
Copy link
Member

hmellor commented Sep 22, 2025

BC for GuidedDecodingParams added in #25422

BC for CLI was already maintained in this PR

The onlt other area we may want to consider is BC in the server API (not sure how we want to handle that)

@russellb
Copy link
Member

I think backwards compatibility in the server API is critical.

https://docs.vllm.ai/en/latest/contributing/deprecation_policy.html#overview

@hmellor
Copy link
Member

hmellor commented Sep 23, 2025

I can make a similar PR ensuring BC for the server API. However I'm not sure how we can warn a user of the deprecated API that it's deprecated?

@russellb
Copy link
Member

I can make a similar PR ensuring BC for the server API. However I'm not sure how we can warn a user of the deprecated API that it's deprecated?

Only in docs for the first step. The next step for the HTTP API would be to turn it off by default, but make an option for turning it back on. Then finally, remove it completely.

russellb added a commit to russellb/vllm that referenced this pull request Sep 23, 2025
This was a mistake introduced by vllm-project#22772. Structured output requests
were not actually working because the format spec was not placed in
the proper new location in the request body.

Signed-off-by: Russell Bryant <[email protected]>
slokesha pushed a commit to slokesha/vllm-gaudi that referenced this pull request Sep 24, 2025
Culprit commit: vllm-project/vllm#22772

---------

Signed-off-by: Agata Dobrzyniewicz <[email protected]>
Signed-off-by: slokesha <[email protected]>
@hmellor
Copy link
Member

hmellor commented Sep 24, 2025

I've just created #25615 to add BC for the server API too

FeiDaLI pushed a commit to FeiDaLI/vllm that referenced this pull request Sep 25, 2025
charlifu pushed a commit to ROCm/vllm that referenced this pull request Sep 25, 2025
…llm-project#22772)

Signed-off-by: Aaron Pham <[email protected]>
Signed-off-by: Harry Mellor <[email protected]>
Co-authored-by: Harry Mellor <[email protected]>
Signed-off-by: charlifu <[email protected]>
dsashidh pushed a commit to dsashidh/pytorch that referenced this pull request Sep 26, 2025
They have been removed in vllm-project/vllm#25117 and vllm-project/vllm#22772, thus failing in trunk at the moment after the latest pin commit update

Pull Request resolved: pytorch#163383
Approved by: https://github.com/wdvr, https://github.com/seemethere, https://github.com/malfet
pytorchbot pushed a commit to pytorch/pytorch that referenced this pull request Sep 30, 2025
They have been removed in vllm-project/vllm#25117 and vllm-project/vllm#22772, thus failing in trunk at the moment after the latest pin commit update

Pull Request resolved: #163383
Approved by: https://github.com/wdvr, https://github.com/seemethere, https://github.com/malfet

(cherry picked from commit a31acf3)
Camyll pushed a commit to pytorch/pytorch that referenced this pull request Sep 30, 2025
Clean up obsoleted vLLM tests (#163383)

They have been removed in vllm-project/vllm#25117 and vllm-project/vllm#22772, thus failing in trunk at the moment after the latest pin commit update

Pull Request resolved: #163383
Approved by: https://github.com/wdvr, https://github.com/seemethere, https://github.com/malfet

(cherry picked from commit a31acf3)

Co-authored-by: Huy Do <[email protected]>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 10, 2025
…llm-project#22772)

Signed-off-by: Aaron Pham <[email protected]>
Signed-off-by: Harry Mellor <[email protected]>
Co-authored-by: Harry Mellor <[email protected]>
Signed-off-by: xuebwang-amd <[email protected]>
choprahetarth pushed a commit to Tandemn-Labs/vllm that referenced this pull request Oct 11, 2025
Angazenn pushed a commit to Angazenn/vllm-ascend that referenced this pull request Oct 21, 2025
…llm-project#2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>


- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 24, 2025
…llm-project#22772)

Signed-off-by: Aaron Pham <[email protected]>
Signed-off-by: Harry Mellor <[email protected]>
Co-authored-by: Harry Mellor <[email protected]>
Signed-off-by: xuebwang-amd <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build documentation Improvements or additions to documentation frontend performance Performance-related issues ready ONLY add when PR is ready to merge/full CI is needed structured-output tool-calling v1

Projects

Status: Done
Status: Done

Development

Successfully merging this pull request may close these issues.

6 participants