<ShubhamAgrawal[7>
My name is Shubham Agrawal. I am currently doing my undergrad in Computer Science.
<ShubhamAgrawal[7>
I think your slack integral broke down cause I can't see any messages here
<ShubhamAgrawal[7>
That I am sending from slack
<ShubhamAgrawal[7>
So, I will ask my question again.
<ShubhamAgrawal[7>
In Adam if I want to use weight decay, then how should I proceed with it? Cause ensmallen ig does not support weight decay with Adam rn.
<ShubhamAgrawal[7>
In ensmallen, for any learning rate-based optimizer such as Adam, SGD, etc. How can we proceed with the group-wise learning rate approach?