generated from fastai/nbdev_template
-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove graph breaks for torch.compile() in padding free branch in DataCollatorForCompletionOnlyLM #2158
Merged
Merged
Remove graph breaks for torch.compile() in padding free branch in DataCollatorForCompletionOnlyLM #2158
Changes from 1 commit
Commits
Show all changes
29 commits
Select commit
Hold shift + click to select a range
4472501
feat: Add info to batch in DataCollatorForCompletionOnlyLM
Abhishek-TAMU 6cfa171
fix: formatting
Abhishek-TAMU a821ce0
feat: Add info to batch in DataCollatorForCompletionOnlyLM
Abhishek-TAMU fb669b6
fix: formatting
Abhishek-TAMU f4b1955
Merge branch 'huggingface:main' into collator_batch
Abhishek-TAMU 1b7c060
Merge branch 'collator_batch' of github.com:Abhishek-TAMU/trl into co…
Abhishek-TAMU c3578f8
Merge branch 'main' into collator_batch
Abhishek-TAMU e83fc8a
fix: max_length_k to int
Abhishek-TAMU 68554b1
fix:Added comments
Abhishek-TAMU 2a7dd47
Merge remote-tracking branch 'trl/main' into collator_batch
Abhishek-TAMU b0a52e2
test cases
Abhishek-TAMU 054a6ef
test cases
Abhishek-TAMU 376ad21
test cases
Abhishek-TAMU 9a08ea3
Merge remote-tracking branch 'trl/main' into collator_batch
Abhishek-TAMU a97045b
feat: Add info to batch in DataCollatorForCompletionOnlyLM
Abhishek-TAMU f31a780
fix: formatting
Abhishek-TAMU 29ba8a3
feat: Add info to batch in DataCollatorForCompletionOnlyLM
Abhishek-TAMU d1441e1
test cases
Abhishek-TAMU d55a6e2
test cases
Abhishek-TAMU 7dccc2d
test cases
Abhishek-TAMU 5e5224e
unit test changes
Abhishek-TAMU 1b434b0
unit test changes
Abhishek-TAMU ef1e304
Merge remote-tracking branch 'trl/main' into collator_batch
Abhishek-TAMU 77894b1
style
qgallouedec 911f60c
Merge branch 'main' into collator_batch
qgallouedec 979f9f0
Merge branch 'main' into collator_batch
qgallouedec cebf936
Merge branch 'main' into collator_batch
qgallouedec ca8e153
add test
qgallouedec 8c27e16
remove test
qgallouedec File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
test cases
Signed-off-by: Abhishek <maurya.abhishek@ibm.com>
- Loading branch information
commit d55a6e280711f5adc05a198ac193bd660f32e91d
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Does the issue only occur with cuda device? In other words can we reproduce on cpu?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Due to usage of
flash_attention_2
it would work only on GPU.