We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b6bf7ba commit a6f2b75Copy full SHA for a6f2b75
vllm/engine/output_processor/multi_step.py
@@ -88,9 +88,15 @@ def process_outputs(self,
88
# TODO: Add support for async if necessary
89
assert not is_async
90
91
+ # Sequences can be in RUNNING or FINISHED_ABORTED state
92
+ # once scheduled, as a sequence is moved to FINSIHED_ABORTED
93
+ # if a client disconnects from the api server.
94
seqs = sequence_group.get_seqs(status=SequenceStatus.RUNNING)
95
+ if seqs is None:
96
+ seqs = sequence_group.get_seqs(
97
+ status=SequenceStatus.FINISHED_ABORTED)
98
- assert seqs, "expected running sequences"
99
+ assert seqs, "Expected RUNNING or FINISHED_ABORTED sequences"
100
assert len(seqs) == 1, (
101
"Beam search not supported in multi-step decoding.")
102
seq = seqs[0]
0 commit comments