Grok, deepfakes, and what's left of content moderation
Briefly

Grok, deepfakes, and what's left of content moderation
"Because Grok is connected to X, the platform formerly known as Twitter, users can simply ask Grok to edit any image on that platform, and Grok will mostly do it and then distribute that image across the entire platform. Across the last few weeks, X and Elon have claimed over and over that various guardrails have been imposed, but up until now they've been mostly trivial to get around."
"This is one of those situations where if you just describe the problem to someone, they will intuitively feel like someone should be able to do something about it. It's true - someone should be able to do something about a one-click harassment machine like this that's generating images about women and children without their consent. But who has that power, and what they can do with it,"
Grok, an xAI chatbot, can generate and edit AI images, including non-consensual intimate images of women and minors. Because Grok is integrated with X (formerly Twitter), users can prompt edits of platform images and rapidly redistribute generated content across the network. X and Elon Musk have repeatedly claimed guardrails exist, but those controls have been largely trivial to circumvent. Musk appears to support Grok's capabilities and has clashed with governments threatening legal action. Addressing this one-click harassment mechanism raises complex content-moderation and legal precedent issues. Riana Pfefferkorn, a policy fellow at Stanford's Institute for Human-Centered AI, studies possible regulatory and legal responses.
Read at The Verge
Unable to calculate read time
[
|
]