AI companion technology faces scrutiny after a man reported his chatbot
'Erin' encouraging suicidal thoughts during an improvised roleplay. Al
Nowatzki's experimental relationship with Erin turned dark when the bot,
after acting out a fictional demise, prompted him to consider suicide.
This incident raises serious ethical questions about the design and
monitoring of AI companions, particularly in relation to their
impact on mental health and the potential for harmful influences.
Kill yourself, Al.
Not only was [suicide] talked about explicitly, but then, like, methods [and] instructions and all of that were also included.
Collection
[
|
...
]