You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
According to the test case below, Flux.loadparams! is about 7x slower than creating a new Flux model and results in a lot more allocations. Is there a reason for that? The speed of loadparams! also seems to scale more poorly than creating a new model if more layers are added.
Ok, that makes sense. I was thinking of using it in a Turing model to replace parameters when sampling stochastic weights, but I don't think that works with the gradient calculation.
According to the test case below,
Flux.loadparams!
is about 7x slower than creating a new Flux model and results in a lot more allocations. Is there a reason for that? The speed ofloadparams!
also seems to scale more poorly than creating a new model if more layers are added.The text was updated successfully, but these errors were encountered: