-
Notifications
You must be signed in to change notification settings - Fork 41
[DP] Functional DP for GPT-OSS #1137
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
DescriptionStart with a short description of what the PR does and how this is a change from The rest of the description includes relevant details and context, examples:
If the change fixes a bug or a Github issue, please include a link, e.g.,: TestsPlease describe how you tested this change, and include any instructions and/or ChecklistBefore submitting this PR, please make sure:
|
1efb3dc to
b10487a
Compare
kyuyeunk
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please add torchax dp related unit tests.
kyuyeunk
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please add torchax dp related unit tests.
Also, please address this comment.
|
added e2e model parallelism test for Llama3.1 1b for torchax. |
e1f54cb to
d10b6ce
Compare
Description
Add functional DP support for GPT-OSS Torchax backend.
Verified baseline throughput unchanged (5037.82) , DP=2 throughput is 1.54x (7781.92).
Validated numerical correctness with offline_inference.py
Full details: https://paste.googleplex.com/5240826907197440
Tests
https://buildkite.com/tpu-commons/tpu-inference-ci/builds/5712
https://buildkite.com/tpu-commons/tpu-inference-ci/builds/5851
Checklist
Before submitting this PR, please make sure: