2842 shaares
32 private links
32 private links
a solid, well-executed paper with a clean idea and good ablations, but limited in ambition by the small scale and synthetic-heavy evaluation. The core insight — that gradient-based memory writing with meta-learned initialization beats forward-only writing — is believable and likely to hold at larger scale, though the computational tradeoff gets harder.