-
Notifications
You must be signed in to change notification settings - Fork 96
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to stack embedding and pass the gradients? #108
Comments
Suppose the output of |
Exactly:
Either concat, or Mean
We need to pass the gradient
and end to end training
… On Feb 23, 2022, at 16:19, Yi-Xuan Xu ***@***.***> wrote:
Suppose the output of N1 and N2 on a sample x is N1(x) and N2(x) separately, what you want is to concatenate their output (i.e., [N1(x), N2(x)]), and pass it to downstream layers, right?
—
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
You are receiving this because you authored the thread.
|
Hi, if you are going to take the mean of outputs from all base estimators, the As to the concatenation, it is somehow weird since all base estimators in the ensemble are doing the same thing, making concatenating their outputs kind of useless. Is there any paper or technical report demonstrating the effectivenss of concatenating outputs of base estimators? |
N1, N2,Nx are different NN models.
We aggregate through concat their embedding output.
BigX = [ X1,….Xn]
and feed into another NN (ie merging).
This extensively used (ie Siamese Network…)
… On Feb 23, 2022, at 23:54, Yi-Xuan Xu ***@***.***> wrote:
Hi, if you are going to take the mean of outputs from all base estimators, the fusion ensemble is exactly what you want.
As to the concatenation, it is somehow weird since all base estimators in the ensemble are doing the same thing, making concatenating their outputs kind of useless. Is there any paper or technical report demonstrating the effectivenss of concatenating outputs of base estimators?
—
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
You are receiving this because you authored the thread.
|
We are NOT dealing with output !!!
Output is kind of useless for End to End training…
We are dealing witht the last embedding.
… On Feb 24, 2022, at 9:41, No Ke ***@***.***> wrote:
N1, N2,Nx are different NN models.
We aggregate through concat their embedding output.
BigX = [ X1,….Xn]
and feed into another NN (ie merging).
This extensively used (ie Siamese Network…)
>> On Feb 23, 2022, at 23:54, Yi-Xuan Xu ***@***.***> wrote:
>>
>
> Hi, if you are going to take the mean of outputs from all base estimators, the fusion ensemble is exactly what you want.
>
> As to the concatenation, it is somehow weird since all base estimators in the ensemble are doing the same thing, making concatenating their outputs kind of useless. Is there any paper or technical report demonstrating the effectivenss of concatenating outputs of base estimators?
>
> —
> Reply to this email directly, view it on GitHub, or unsubscribe.
> Triage notifications on the go with GitHub Mobile for iOS or Android.
> You are receiving this because you authored the thread.
|
Thanks for your kind explanation. Heterogeneous ensemble is not supported yet, since we have not come up with a succinct way on setting different optimizers for different base estimators 😢. |
Sure.
At 1st version,
Maybe, we can use same optimizer, scheduler for the ensemble model
Goal is to have a one liner for easy
Ensemble End to End.
… On Feb 24, 2022, at 10:04, Yi-Xuan Xu ***@***.***> wrote:
Thanks for your kind explanation. Heterogeneous ensemble is not supported yet, since we have not come up with a succinct way on setting different optimizers for different base estimators 😢.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
You are receiving this because you authored the thread.
|
Kind of busy these days, will appreciate a PR very much ;-) |
Have a 2 neural nets N1, N2,
want to stack their output embedding layer.
How to do this ?
The text was updated successfully, but these errors were encountered: