+1 for a Mechanism Design/Aligning Incentives tag. I think “incentive design” would be a good name for this category. This would encompass material on specification gaming, tampering, impact measures, etc. Including specific examples of misaligned incentives under this umbrella seems fine as well.
Is the “aligning incentives” tag you are interested in something AI specific or should it apply to general human institutions / social systems? I could see a case for either, but that impacts what tag names we should use.
+1 for a Mechanism Design/Aligning Incentives tag. I think “incentive design” would be a good name for this category. This would encompass material on specification gaming, tampering, impact measures, etc. Including specific examples of misaligned incentives under this umbrella seems fine as well.
Is the “aligning incentives” tag you are interested in something AI specific or should it apply to general human institutions / social systems? I could see a case for either, but that impacts what tag names we should use.
I was thinking of an AI specific tag, it seems a bit too broad otherwise.