Oh, hmm, this seems like a bug on our side. I definitely set up a redirect a while ago that should make those links work. My guess is something broke in the last few months.
LW confusingly replaces the link to www.alignmentforum.org given in Markdown comment source text with a link to www.lesswrong.com when displaying the comment on LW.
A framing I wrote up for a debate about “alignment tax”:
“Alignment isn’t solved” regimes:
Nobody knows how to make an AI which is {safe, general, and broadly superhuman}, with any non-astronomical amount of compute
We know how to make an aligned AGI with 2 to 25 OOMs more compute than making an unaligned one
“Alignment tax” regimes:
We can make an aligned AGI, but it requires a compute overhead in the range 1% − 100x. Furthermore, the situation remains multipolar and competitive for a while.
The alignment tax is <0.001%, so it’s not a concern.
The leading coalition is further ahead than the alignment tax amount, and can and will execute a pivotal act, thus ending the risk period and rendering the alignment tax irrelevant.
A person whose mainline is {1a --> 1b --> 2b or 2c} might say “alignment is unsolved, solving it mostly a discrete thing, and alignment taxes and multipolar incentives aren’t central”
Whereas someone who thinks we’re already in 2a might say “alignment isn’t hard, the problem is incentives and competitiveness”
Someone whose mainline is {1a --> 2a} might say “We need to both ‘solve alignment at all’ AND either get the tax to be really low or do coordination. Both are hard, and both are necessary.”
It’s sad that agentfoundations.org links no longer work, leading to broken links in many decision theory posts (e.g. here and here)
Oh, hmm, this seems like a bug on our side. I definitely set up a redirect a while ago that should make those links work. My guess is something broke in the last few months.
Thanks for the heads up. Example broken link (
https://agentfoundations.org/item?id=32
), currently redirects to brokenhttps://www.alignmentforum.org/item?id=32
, should redirect further tohttps://www.alignmentforum.org/posts/5bd75cc58225bf0670374e7d/exploiting-edt
(Exploiting EDT[1]), archive.today snapshot.Edit 14 Oct: It works now, even for links to comments, thanks LW team!
LW confusingly replaces the link to
www.alignmentforum.org
given in Markdown comment source text with a link towww.lesswrong.com
when displaying the comment on LW.A framing I wrote up for a debate about “alignment tax”:
“Alignment isn’t solved” regimes:
Nobody knows how to make an AI which is {safe, general, and broadly superhuman}, with any non-astronomical amount of compute
We know how to make an aligned AGI with 2 to 25 OOMs more compute than making an unaligned one
“Alignment tax” regimes:
We can make an aligned AGI, but it requires a compute overhead in the range 1% − 100x. Furthermore, the situation remains multipolar and competitive for a while.
The alignment tax is <0.001%, so it’s not a concern.
The leading coalition is further ahead than the alignment tax amount, and can and will execute a pivotal act, thus ending the risk period and rendering the alignment tax irrelevant.
A person whose mainline is {1a --> 1b --> 2b or 2c} might say “alignment is unsolved, solving it mostly a discrete thing, and alignment taxes and multipolar incentives aren’t central”
Whereas someone who thinks we’re already in 2a might say “alignment isn’t hard, the problem is incentives and competitiveness”
Someone whose mainline is {1a --> 2a} might say “We need to both ‘solve alignment at all’ AND either get the tax to be really low or do coordination. Both are hard, and both are necessary.”
Results on logarithmic utility and stock market leverage: https://www.lesswrong.com/posts/DMxe4XKXnjyMEAAGw/the-geometric-expectation?commentId=yuRie8APN8ibFmRJD