Maybe if you solve for equilibrium you get that after releasing the tool, the tool is defeated reasonably quickly?
I believe it’s already known that running the text through another (possibly smaller and cheaper) LLM to reword it can remove the watermarking. So for catching cheaters it’s only a tiny bit stronger than searching for “as a large language model” in the text.
I believe it’s already known that running the text through another (possibly smaller and cheaper) LLM to reword it can remove the watermarking. So for catching cheaters it’s only a tiny bit stronger than searching for “as a large language model” in the text.