Stanford MLSys Seminars is where talks from the Hazy Research group at stanford get posted, and their work has been some of the most eye-catching for me in the past two years. In particular, the S4 sequence model seems to me to represent a major capability bump in next-step-after-transformers models, due to its unusually stable learning. I might just be taken in by a shiny toy, but S4 is the next thing I’m going to play with capabilities wise. https://www.youtube.com/c/StanfordMLSysSeminars
Stanford MLSys Seminars is where talks from the Hazy Research group at stanford get posted, and their work has been some of the most eye-catching for me in the past two years. In particular, the S4 sequence model seems to me to represent a major capability bump in next-step-after-transformers models, due to its unusually stable learning. I might just be taken in by a shiny toy, but S4 is the next thing I’m going to play with capabilities wise. https://www.youtube.com/c/StanfordMLSysSeminars