Skip to content

Conversation

Superjomn
Copy link
Collaborator

@Superjomn Superjomn commented May 23, 2025

Description

  1. Eliminate the pytorch_backend_config and flatten all PyTorchConfig knobs into TorchLlmArgs
  2. Update all the usages accordingly

@Superjomn Superjomn force-pushed the flatten-pytorch-config branch 3 times, most recently from 678236c to cc4863c Compare May 23, 2025 06:31
@Superjomn Superjomn requested a review from a team as a code owner May 23, 2025 06:31
@Superjomn Superjomn requested review from yilin-void and Funatiq May 23, 2025 06:31
@Superjomn Superjomn force-pushed the flatten-pytorch-config branch 3 times, most recently from c7573da to 1ec59ec Compare May 23, 2025 08:37
@Superjomn Superjomn requested a review from QiJune May 23, 2025 08:40
@Superjomn Superjomn force-pushed the flatten-pytorch-config branch 2 times, most recently from 19f20a3 to c3d0f10 Compare May 23, 2025 09:51
@Superjomn
Copy link
Collaborator Author

/bot run --add-multi-gpu-test --disable-fail-fast

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6285 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6285 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #4593 completed with status: 'FAILURE'

@Superjomn Superjomn force-pushed the flatten-pytorch-config branch from c3d0f10 to 1115607 Compare May 25, 2025 08:18
@Superjomn Superjomn requested a review from a team as a code owner May 25, 2025 08:18
@Superjomn
Copy link
Collaborator Author

/bot run --add-multi-gpu-test --disable-fail-fast

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6379 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6379 [ run ] completed with state FAILURE
/LLM/main/L0_MergeRequest_PR pipeline #4661 completed with status: 'FAILURE'

@Superjomn
Copy link
Collaborator Author

/bot run --add-multi-gpu-test --disable-fail-fast

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6418 [ run ] triggered by Bot

@Superjomn Superjomn force-pushed the flatten-pytorch-config branch 2 times, most recently from 7755876 to a9f24cb Compare May 26, 2025 05:29
@Superjomn
Copy link
Collaborator Author

/bot run --add-multi-gpu-test --disable-fail-fast

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6425 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6418 [ run ] completed with state ABORTED

@Superjomn Superjomn force-pushed the flatten-pytorch-config branch from a9f24cb to 6393176 Compare May 27, 2025 01:57
@tensorrt-cicd
Copy link
Collaborator

PR_Github #6582 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6582 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #4814 completed with status: 'FAILURE'

@Superjomn Superjomn force-pushed the flatten-pytorch-config branch from 2624ad4 to 869e6c8 Compare May 27, 2025 11:21
@Superjomn Superjomn requested review from a team as code owners May 27, 2025 11:21
@Superjomn
Copy link
Collaborator Author

/bot run --disable-fail-fast

@Superjomn Superjomn enabled auto-merge (squash) May 27, 2025 11:22
@tensorrt-cicd
Copy link
Collaborator

PR_Github #6630 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6630 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #4845 completed with status: 'SUCCESS'

Superjomn added 6 commits May 28, 2025 10:19
Signed-off-by: Superjomn <[email protected]>
update usages

Signed-off-by: Superjomn <[email protected]>
Signed-off-by: Superjomn <[email protected]>
Signed-off-by: Superjomn <[email protected]>
Signed-off-by: Superjomn <[email protected]>
@Superjomn Superjomn force-pushed the flatten-pytorch-config branch from 869e6c8 to bee40fa Compare May 28, 2025 10:22
@Superjomn
Copy link
Collaborator Author

/bot reuse-pipeline

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6763 [ reuse-pipeline ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #6763 [ reuse-pipeline ] completed with state SUCCESS
Reusing PR_Github #6630 for commit bee40fa

@Superjomn Superjomn merged commit 5506f60 into NVIDIA:main May 28, 2025
3 checks passed
@Superjomn Superjomn deleted the flatten-pytorch-config branch May 28, 2025 13:42
YOUR_DATA_PATH=<your dataset file following the format>

cat >./extra-llm-api-config.yml<<EOF
pytorch_backend_config:
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we should throw a "deprecated" message when the old config type is detected. Currently old configs are still accepted, but all the fields under pytorch_backend_config are ignored, thus none of the configs are taking effect. This causes confusion to users (like myself).

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think this is a very fair request. @Superjomn

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

10 participants