Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

OOM when using encode for token_embeddings #1813

Open
bodin-e opened this issue Jan 19, 2023 · 0 comments
Open

OOM when using encode for token_embeddings #1813

bodin-e opened this issue Jan 19, 2023 · 0 comments

Comments

@bodin-e
Copy link

bodin-e commented Jan 19, 2023

I am running into OOM when applying encode for token embeddings on a large dataset.

Currently the solution to resolve OOM using the encode method (see #522 and #487) is only applicable to sentence embeddings, not token embeddings.

I have resolved the issue by generalizing the previous solution to also be applicable to token embeddings via the use of an added 'move_to_cpu' flag. Is their an alternative approach that I have missed? If not, and you agree with the changes, feel free to merge #1812.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant