
We understand nothing, we just made it colorful and called it science
A technical visualization post from WelchLabs showing the internal mechanisms of a neural network with mathematical wave equations and 3D surface plots. The post satirically claims this is 'the most complex model we actually understand,' with a comment joking that 95% of ML researchers quit before achieving grokking. It's a meta-joke about how little we actually understand about deep learning internals despite appearing to decode them.
Extracted text:
cos(113π x), sin(113π x), SPARSE LINEAR MODE, EMBED, SPARSE PROBE, ATTENTION, cos(113π y), sin(113π y), A₄ cos(113π + φ₃), A₃ cos(113π + φ₃), A₄ cos(113π y + φ₃), A₁ cos(113π x + φ₁), The most complex model we actually und..., @WelchLabs 626k views 1 mo ago ...more, 26k, Save, Comments 1.3k, 95% of ML Researches give up on training one epoch before grokking occurs.