Skip to content

how to use V1 engine in vllm mode #1586

@ZCSlqc

Description

@ZCSlqc

I used A10 GPU to run CosyVoice2-0.5B in VLLM mode, but encountered the message: "--enable-prompt-embeds is not supported by the V1 Engine. Falling back to V0". When I commented out this line in the  load_vllm()  function of  cosyvoice/cli/model.py , the model initialization failed this time.

I would appreciate your help with this issue.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions