Why do a lot of the “External reviews of “Alignment faking in large language models”” read like they were also written, or edited, by LLMs?
Are people expected to take “reviews” done seemingly pro forma at face value?
Why do a lot of the “External reviews of “Alignment faking in large language models”” read like they were also written, or edited, by LLMs?
Are people expected to take “reviews” done seemingly pro forma at face value?