Skip to content

Packed tokenizer #1473

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Apr 12, 2025
Merged

Packed tokenizer #1473

merged 1 commit into from
Apr 12, 2025

Conversation

AngledLuffa
Copy link
Collaborator

@AngledLuffa AngledLuffa commented Apr 5, 2025

Use a PackedSequence in the tokenizer. Somehow it's substantially slower...

however, it would address the tokenization not being consistent based on batch size:

#1472

@AngledLuffa AngledLuffa merged commit 4433e83 into dev Apr 12, 2025
1 check passed
@AngledLuffa AngledLuffa deleted the packed_tokenizer branch April 12, 2025 00:08
Sort in the other direction means we don't need to use enforce_sorted=False

Things are faster without the packed sequences, unfortunately, but they wind up having unstable results:

#1472
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant