We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 2b16874 commit 61fbbd7Copy full SHA for 61fbbd7
docs/changelogs/v3.3.3.md
@@ -4,5 +4,7 @@
4
5
* Implement `Grams` optimizer. (#317, #318)
6
* [Grams: Gradient Descent with Adaptive Momentum Scaling](https://arxiv.org/abs/2412.17107)
7
-* Support `stable_adamw` variant for `ADOPT` and `AdEMAMix` optimizer. (#319)
+* Support `stable_adamw` variant for `ADOPT` and `AdEMAMix` optimizer. (#320)
8
* `optimizer = ADOPT(model.parameters(), ..., stable_adamw=True)`
9
+* Implement an experimental optimizer `Ranger25` (not tested). (#320)
10
+ * mixing `ADOPT + AdEMAMix + StableAdamW + Cautious + RAdam` optimizers.
0 commit comments