Prudential reasons for designing entitled chatbots: How robot "rights" can improve human well-being

设计拥有自主权的聊天机器人的审慎理由:机器人“权利”如何改善人类福祉

阅读:1

Abstract

Can robots or chatbots be moral patients? The question of robot rights is often linked to moral reasons like precautionary principles or the ability to suffer. We argue that we have prudential reasons for building robots that can at least hold us accountable (criticize us etc.) and that we have prudential reasons to build robots that can demand that we treat them with respect. This proposal aims to add nuance to the robot rights debate by answering a key question: Why should we want to build robots that could have rights in the first place? We argue that some degree of accountability in our social relationships contributes to our well-being and flourishing. The normativity ascribed to robots will increase their social and non-social functionalities from action coordination to more meaningful relationships. Having a robot that has a certain "standing" to hold us accountable can improve our epistemic standing and satisfy our desire for recognition.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。