Skip to content

Conversation

yunfeng-scale
Copy link
Contributor

Reverts #380

@yunfeng-scale yunfeng-scale requested a review from a team November 17, 2023 01:43
# Based on config here: https://huggingface.co/TIGER-Lab/MAmmoTH-Coder-7B/blob/main/config.json#L12
# Can also see 13B, 34B there too
"codellama": {"max_model_len": 16384, "max_num_batched_tokens": 16384},
"code-llama": {"max_model_len": 16384, "max_num_batched_tokens": 16384},
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

quick question on this Yunfeng: I wanted the models to be named codellama to be in keeping with the way that it's formatted for finetuning, why are we switching?

@ian-scale ian-scale self-requested a review November 17, 2023 01:49
@yunfeng-scale yunfeng-scale enabled auto-merge (squash) November 17, 2023 02:04
@yunfeng-scale yunfeng-scale merged commit d478ee5 into main Nov 17, 2023
@yunfeng-scale yunfeng-scale deleted the revert-380-ss/two-quick-vllm-bugs branch November 17, 2023 02:05
@yunfeng-scale yunfeng-scale mentioned this pull request Mar 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants