Skip to content

Conversation

@krishnaraj36
Copy link
Contributor

@krishnaraj36 krishnaraj36 commented May 7, 2024

Enhanced the opencl thread limit and improved the gpu schedules for opencl targets.
It improves decode performance 20 % for few set of MLC llm models.

LLM model --- baseline --- improved
gemma-2b-it --- 22.4 tok/sec --- 28.2 tok/sec
Qwen-7b-chat --- 11 tok/sec --- 11.8 tok/sec

Enhanced the opencl thread limit and improved the gpu schedules
for opencl targets.
It improves decode performance 20 % for few set of models.
@krishnaraj36
Copy link
Contributor Author

@srkreddy1238 @tqchen : Please take a look to this PR and let me know your advise.

@tqchen
Copy link
Member

tqchen commented May 9, 2024

cc @mengshyu would be nice to confirm metal improvement and see if we want it for webgpu/metal

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants