This week we spotlight the thirteenth risk framework included in the AI Risk Repository:
This paper presents guidance for evaluating the broad social impacts of generative AI systems across two overarching categories: what can be evaluated in relation to the technical ‘base’ system and what can be evaluated among people and society.
For the base system, the framework defines 6 categories of social impact for evaluation: bias, stereotypes, and representational harms; cultural values and sensitive content; disparate performance; privacy and data protection; financial costs; environmental costs; and data and content moderation labor costs. For the broader societal context, the framework defines 5 categories of social impact for evaluation: trustworthiness and autonomy; inequality, marginalization, and violence; concentration and authority; labor and creativity; and ecosystem and environment.
This summary highlights a paper included in the MIT AI Risk Repository. We did not author the paper and credit goes to the authors. For the full details, please refer to the original publication: https://arxiv.org/pdf/2306.05949.