Update: I have now cross-referenced every single email for accuracy, cleaned up and clarified the thread structure, and added subject lines and date stamps wherever they were available. I now feel comfortable with people quoting anything in here without checking the original source (unless you are trying to understand the exact thread structure of who was CC’d and when, which was a bit harder to compress into a linear format).
(For anyone curious, the AI transcription and compilation made one single error, which is that it fixed a typo in one of Sam’s messages from “We did this is a way” to “We did this in a way”. Honestly, my guess is any non-AI effort would have had a substantially higher error rate, which was a small update for me on the reliability of AI for something like this, and also makes the handwringing about whether it is OK post something like this feel kind of dumb. It also accidentally omitted one email with a weird thread structure.)
Thanks for not only doing this but noting the accuracy of the unchecked transcript, it’s always hard work to build a mental model of how good LLM tools are at what stuff.
That requires interpretation, which can introduce unintended editorializing. If you spotted the intent, the rest of the audience can as well. (And if the audience is confused about intent, the original recipients may have been as well.)
I personally would include these sorts of notes about typos if I was writing my own thoughts about the original content, or if I was sharing a piece of it for a specific purpose. I take the intent of this post to be more of a form of accessible archiving.
Update: I have now cross-referenced every single email for accuracy, cleaned up and clarified the thread structure, and added subject lines and date stamps wherever they were available. I now feel comfortable with people quoting anything in here without checking the original source (unless you are trying to understand the exact thread structure of who was CC’d and when, which was a bit harder to compress into a linear format).
(For anyone curious, the AI transcription and compilation made one single error, which is that it fixed a typo in one of Sam’s messages from “We did this is a way” to “We did this in a way”. Honestly, my guess is any non-AI effort would have had a substantially higher error rate, which was a small update for me on the reliability of AI for something like this, and also makes the handwringing about whether it is OK post something like this feel kind of dumb. It also accidentally omitted one email with a weird thread structure.)
Thanks for not only doing this but noting the accuracy of the unchecked transcript, it’s always hard work to build a mental model of how good LLM tools are at what stuff.
Huzzah for assembling conversations! With this proof of concept, I wonder how easy it will be to deploy inside of LessWrong here.
I wonder if it would be a good idea to put editor’s notes after likely typos, like:
That requires interpretation, which can introduce unintended editorializing. If you spotted the intent, the rest of the audience can as well. (And if the audience is confused about intent, the original recipients may have been as well.)
I personally would include these sorts of notes about typos if I was writing my own thoughts about the original content, or if I was sharing a piece of it for a specific purpose. I take the intent of this post to be more of a form of accessible archiving.