AdamW optimizer with optional weight decay.
- Described: https://paperswithcode.com/method/adamw
- Paper: https://arxiv.org/abs/1711.05101v3
const
AdamW = (params:
Tensor[], lr: number = 0.001, b1: number = 0.9, b2: number = 0.999, eps: number = 0.00000001, weight_decay: number = 0.01) => unknown