Introducing WW-PGD: A Cutting-Edge Add-On for Optimizer Enhancement ๐
Discover the latest release of WW-PGD, a PyTorch add-on designed to supercharge your model training by integrating epoch-boundary spectral projections with standard optimizers. Unleash optimized performance and detailed spectral control in your deep learning workflows!
Announcing: ๐ช๐ช-๐ฃ๐๐ โ ๐ช๐ฒ๐ถ๐ด๐ต๐๐ช๐ฎ๐๐ฐ๐ต๐ฒ๐ฟ ๐ฃ๐ฟ๐ผ๐ท๐ฒ๐ฐ๐๐ฒ๐ฑ ๐๐ฟ๐ฎ๐ฑ๐ถ๐ฒ๐ป๐ ๐๐ฒ๐๐ฐ๐ฒ๐ป๐ ๐
I’m thrilled to announce the release of WW-PGDโa novel PyTorch add-on designed to empower your deep learning optimization process. This small yet powerful tool wraps around standard optimizers like SGD, Adam, and AdamW, incorporating an epoch-boundary spectral projection powered by WeightWatcher diagnostics.
๐ Elevator Pitch
WW-PGD doesn’t just optimize; it strategically nudges each layer towards the Exact Renormalization Group (ERG) critical manifold during training. This approach ensures that you’re aiming for the right optimization targets right from the get-go, rather than relying on post-hoc diagnostics.
๐ Theory in Short
- HTSR Critical Condition: ฮฑ โ 2
- SETOL ERG Condition: trace-log(ฮป) over the spectral tail = 0
By making these conditions explicit optimization goals, WW-PGD brings a new level of precision to layer management during training.
โ๏ธ How It Works
Here’s a quick overview of the mechanics:
- Runs WeightWatcher (ww) at Epoch Boundaries: At the end of each epoch, WW-PGD evaluates the model’s weight distribution.
- Identifies the Spectral Tail: Utilizes layer quality metrics from ww to determine which portion of the weight distribution is the spectral tail.
- Optimal Tail Guess Selection: It selects an optimal guess for the tail at each epoch.
- Applies Projected Gradient Descent Update: Uses a stable, Cayley-like Proximal step to update the layerโs spectral density.
- Retracts to Satisfy SETOL ERG Condition: Ensures that updates adhere to the spectral constraints.
- Blends Projected Weights Back In: Incorporates a "warmup" + ramping process to avoid instability early on.
In essence, WW-PGD provides a mechanism to project the optimizer’s results onto the ERG critical manifold, enhancing efficiency in spectral constraint optimization.
๐ Scope (Important)
This initial public release is tailored for training small models from scratch, and is not yet optimized for large-scale fine-tuning tasks. Consider it a proof of concept, with ongoing tests extending to:
- 3-layer MLPs (MNIST / FashionMNIST)
- nano-GPT-style small Transformer models
Future work is dedicated to adapting larger architectures and fine-tuning workflows.
๐ Early Results (FashionMNIST, 35 Epochs, Mean ยฑ Std)
The initial tests yield intriguing results:
- Plain Test: Baseline 98.05% ยฑ 0.13 vs WW-PGD 97.99% ยฑ 0.17
- Augmented Test: Baseline 96.24% ยฑ 0.17 vs WW-PGD 96.23% ยฑ 0.20
This indicates that while accuracy is nearly neutral at this scale, WW-PGD offers a significant advantage with a spectral control knob and comprehensive per-epoch tuning.
๐ฅ Repo & QuickStart
- ๐งฉ Repo: GitHub Repository
- ๐ QuickStart (with MLP3+FashionMNIST example): QuickStart Guide
- ๐ More Info: WeightWatcher
If you’re experimenting with training and optimization on your models, or looking for a data-free spectral health monitor + projection step, your feedback is invaluable. Join us in exploring other optimizers or small Transformer setups!
๐ฌ Community Engagement
Join the WeightWatcher Community on Discord to share insights and learn from fellow developers: Discord Invitation
A special thanks to Hari Kishan Prakash for his invaluable contributions to this project!
If you have any questions or need assistance with AI, feel free to reach out. Letโs talk! #talkToChuck