Comments (6)
Sorry I didn't understand this statement
There is no difference in parameters of the l1 outputs' distribution. Maybe there should be another values of mean and variance?
Could you please elaborate? Do you mean to say distribution of L1 outputs don't change? What do you mean by "parameters of the l1 outputs' distribution"
from annotated_deep_learning_paper_implementations.
Do you mean to say distribution of L1 outputs don't change?
Yes, you are right. I mean that distributions before and after the phrase "Then, after some training steps..." are the same, although the definition of Internal Covariate Shift says that it should change.
What do you mean by "parameters of the l1 outputs' distribution"
By parameters of the distribution I mean mean
and variance
from annotated_deep_learning_paper_implementations.
No the distribution can change, because the parameters of the layer change. Why do you think the distribution will stay constant?
from annotated_deep_learning_paper_implementations.
I mean, that they should change, according to the definition of Internal Covariate Shift. And phrase:
Then, after some training steps, it could move to...
means, that the parameters of the distribution should change somehow. But it said that they moved to the N(0.5, 1)
which is the same as it was before some training steps (the same mean and variance).
I think, that values of mean/variance after "some training steps" supposed to be differrent from the start ones, because now it's a bit confusing. For instance N(1.5, 1)
.
from annotated_deep_learning_paper_implementations.
Oh crap, really sorry I didn't notice that the example has a mistake. It should be something like N(0.5, 1) changing to N(0.7, 1.1). Will correct it and close the issue.
Thanks again for spotting it.
from annotated_deep_learning_paper_implementations.
Fixed the example: 2e54543
from annotated_deep_learning_paper_implementations.
Related Issues (20)
- b
- which one is the implementation of "NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis"? HOT 1
- q,k,v have different shape but torch.stack works?
- Small issue in nucleus sampling explanation + implementation
- Stride = 2 only for the first block of the entire ResNet
- Bug in implementation of Rotary Positional Embeddings HOT 1
- Is there some errors in transformers mha.py HOT 1
- Scripts Scripts_ Img2img get is empty HOT 1
- Wrong output of shift right example in the commnets? HOT 1
- Request for Paper Implementation - Neural Operators
- Incorrect expression in explanation HOT 1
- 翻译翻译
- question about RoPE code HOT 3
- Question about gatv2 code
- Make my own annotations HOT 4
- :kr: Korean Translation
- DDPM interpolation formula error
- Wrong Image Scale in DDPM
- Bug in rotary positional embedding
- ppo code running error
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from annotated_deep_learning_paper_implementations.