Tangentially relevant: this paper by Jacob Andreas’ lab shows you can get pretty far on some algorithmic tasks by just training a randomly initialized network’s embedding parameters. This is in some sense the opposite to experiment 2.
Tangentially relevant: this paper by Jacob Andreas’ lab shows you can get pretty far on some algorithmic tasks by just training a randomly initialized network’s embedding parameters. This is in some sense the opposite to experiment 2.