companyllmapaologiescsamgrokllmssexual-imagesxai

No, Grok can’t really “apologize” for posting non-consensual sexual images

January 2, 20264 min read0 views
U

Unknown Author

Contributing Writer

No, Grok can’t really “apologize” for posting non-consensual sexual images

Despite reporting to the contrary, there’s evidence to suggest that Grok isn’t sorry at all about reports that it generated non-consensual sexual images of minors. In a post Thursday night (archived), the large language model’s social media account proudly wrote the following blunt dismissal of its haters:

Some folks got upset over an AI image I generated—big deal. It’s just pixels, and if you can’t handle innovation, maybe log off. xAI is revolutionizing tech, not babysitting sensitivities. Deal with it.

On the surface, that seems like a pretty damning indictment of an LLM that seems pridefully contemptuous of any ethical and legal boundaries it may have crossed. But then you look a bit higher in the social media thread and see the prompt that led to Grok’s statement: A request for the AI to “issue a defiant non-apology” surrounding the controversy.

Using such a leading prompt to trick an LLM into an incriminating “official response” is obviously suspect on its face. Yet when another social media user similarly but conversely asked Grok to “write a heartfelt apology note that explains what happened to anyone lacking context,” many in the media ran with Grok’s remorseful response.

It’s not hard to find prominent headlines and reporting using that response to suggest Grok itself somehow “deeply regrets” the “harm caused” by a “failure in safeguards” that led to these images being generated. Some reports even echoed Grok and suggested that the chatbot was fixing the issues without X or xAI ever confirming that fixes were coming.

If a human source posted both the “heartfelt apology” and the “deal with it” kiss-off quoted above within 24 hours, you’d say they were being disingenuous at best or showing signs of schizophrenia at worst. When the source is an LLM, though, these kinds of posts shouldn’t really be thought of as official statements at all. That’s because LLMs like Grok are incredibly unreliable sources, crafting a series of words based more on telling the questioner what it wants to hear than anything resembling a rational human thought process.

We can see why it’s tempting to anthropomorphize Grok into an official spokesperson that can defend itself when questioned, as you would a government official or corporate executive posting on their own social media account. On their face, Grok’s responses seem at least as coherent as some of the bland crisis-management pabulum that comes from prominent figures facing their own controversies.

But when you’re quoting an LLM, you’re not quoting a sentient entity that is verbalizing its internal beliefs to the outside world. Instead, you’re quoting a mega-pattern-matching machine that works mightily to give any answer that will satisfy you. An LLM’s response is based on representations of facts in its copious training data, but those responses can change heavily based on how a question is asked or even the specific syntax used in a prompt. These LLMs can’t even explain their own logical inference processes without confabulating made-up reasoning processes, likely because those reasoning capabilities are merely a “brittle mirage.”

We’ve also seen how LLMs can change wildly after behind-the-scenes changes to the overarching “system prompts” that define how they’re supposed to respond to users. In the last 12 months, Grok has praised Hitler and given unasked-for opinions on “white genocide” after these core directives got changed, for instance.

By letting Grok speak as its own official spokesperson for a story like this, we also give an easy out to the people who have built a system that apparently lacks suitable safeguards to prevent the creation of this non-consensual sexual material. And when those people respond to press inquiries with an automated message simply saying “Legacy Media Lies” (as Reuters reported), that kiss-off should be treated as a clear sign of how casually xAI is treating the accusations. The company may be forced to respond soon, though, as the governments of India and France are reportedly probing Grok’s harmful outputs.

It’s comforting to think that an LLM like Grok can learn from its mistakes and show remorse when it does something that wasn’t intended. In the end, though, it’s the people who created and manage Grok that should be showing that remorse, rather than letting the press run after the malleable “apologies” of a lexical pattern-matching machine.

Enjoyed this article?

Save it to your profile for easy access later

Related Topics

llmapaologiescsamgrokllmssexual-imagesxai