You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I mean, they are inferable by the reader, but it looks like they are intended to be visible in the finished product and getting eaten somewhere along the publication tool chain?
I would also say that this tutorial reads a little bit bouncier between levels of technical explanation than one might hope. I think the call out to the reader to supply their own loss function is more of a math textbook kind of exercise left to the reader than something that belongs in explanatory documentation. The bit about BatchNormalization being dependent on batch size and therefore not properly vectorizable also seems like a theoretical concern that should be left out or handled differently.
I mean, they are inferable by the reader, but it looks like they are intended to be visible in the finished product and getting eaten somewhere along the publication tool chain?
I would also say that this tutorial reads a little bit bouncier between levels of technical explanation than one might hope. I think the call out to the reader to supply their own loss function is more of a math textbook kind of exercise left to the reader than something that belongs in explanatory documentation. The bit about BatchNormalization being dependent on batch size and therefore not properly vectorizable also seems like a theoretical concern that should be left out or handled differently.
flax/docs/guides/flax_fundamentals/state_params.rst
Line 12 in 514c111
The text was updated successfully, but these errors were encountered: