-
Notifications
You must be signed in to change notification settings - Fork 16
Add RoBERTa FP8 support with refactoring #72
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Merged
Changes from 3 commits
Commits
Show all changes
23 commits
Select commit
Hold shift + click to select a range
4d57dd0
initial encoder refactoring (wip)
andrea-fasoli 6de2a31
fp8 encoder support
andrea-fasoli 0c6a36b
Update detection of RoBERTa architecture
andrea-fasoli abc01e5
Remove commented decoder args
andrea-fasoli 0c84b8b
Make verbose a flag
andrea-fasoli 66780e3
Remove TODO in FP8 quantization argument
andrea-fasoli bde2c60
Remove decoder arguments from argument validation
andrea-fasoli e8c2cd5
Update argument help
andrea-fasoli e66cd1c
Update padding explanation
andrea-fasoli e5555c8
Update linear config message for FP8
andrea-fasoli af3d681
raise error for default_dtype + quantization
andrea-fasoli d8b73ee
Update printouts
andrea-fasoli 55f19d8
Update determinism docstring
andrea-fasoli 2d8b88d
Fix typos
andrea-fasoli 429c57f
Update rank-based printouts
andrea-fasoli 041f39a
Remove superseeded roberta.py script
andrea-fasoli 0dfa472
Gate post processing to rank 0 only
andrea-fasoli bca7a39
Rename encoder inference entry point script
andrea-fasoli fb3f224
merge from upstream/main
andrea-fasoli e989a23
Move batch to correct device at eval
andrea-fasoli 0cd0d7b
Add 16b forced casting
andrea-fasoli 18f4230
Reinstate local_size in aiu_setup (for future use)
andrea-fasoli 2fc4c12
Add notes about 384 default max_prompt_length
andrea-fasoli File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we'll eventually need local_rank (when multi-aiu moves to multi-node multi-aiu)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
good point to keep in mind. Should we add it back at that time or now?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
generally
torchrun
already provides it, so we might as well keep it even if it's the same as RANK for now. Some models/algorithms might even expect to use that instead of rank, so it's good to have bothThere was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
OK,
local_rank
andlocal_size
are not in use right now but we keep them in the setup for future needs