-
Notifications
You must be signed in to change notification settings - Fork 287
Do not use an explicit loop in unflatten #573
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
...to avoid many small GPU <-> CPU transfers.
|
@explosion-bot please test_gpu |
URL: https://buildkite.com/explosion-ai/thinc-gpu-test-suite/builds/15 |
298489f to
ba295da
Compare
ba295da to
fcc572c
Compare
|
@danieldk : looks good! Feel free to merge in once you're done with your benchmarks and everything seems OK :-) |
|
Just a status update: I want to benchmark this a little more. |
|
I did more benchmarks and this change seems to make training slightly slower (2-3%) and inference slightly faster by the same amount. Since a declarative style is nicer in general and since this should scale better (the looping over the array will get more expensive with longer lengths arrays), I'll merge this. |
...to avoid many small GPU <-> CPU transfers.
I still have to benchmark this change to see if this leads to a meaningful training time difference.