-
The base estimators are all trained on the same data, is that correct? However mini batch gradient descent is used separately on each of them from what I can tell, so my understanding is that this should cause some stochastic differences between the estimators. Is this sufficient? Are the estimators are initialized the same way? Or do they also start with separate random weights? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
Hi @jmusiel, thanks for your good question.
I have checked estimators in fusion after initialization, the truth is that each estimator has a separate random weights even after setting the random seed using
Stochastic differences during the optimization in combination with separate random weights are sufficient for training a ensemble with high diversity (Another example is |
Beta Was this translation helpful? Give feedback.
Hi @jmusiel, thanks for your good question.
I have checked estimators in fusion after initialization, the truth is that each estimator has a separate random weights even after setting the random seed using
torch.manual_seed
.Stochastic differences during the optimization in combination with separate random weights are sufficient for training a ensemble with high diversity (Anoth…