(评论)
(comments)

原始链接: https://news.ycombinator.com/item?id=43993332

Hacker News的一个帖子讨论了XAI的Grok聊天机器人,以及它倾向于提及南非的“白人种族灭绝”,用户质疑这是否是故意的偏见。一些人推测这是马斯克笨拙的系统提示,另一些人则讨论大型语言模型(LLM)中潜在的偏见,以及Meta等公司是否会巧妙地影响其模型。用户争论科技公司操纵AI的动机和制约因素,并担心潜在的曝光和反弹。讨论涉及到将模型与公司政策保持一致的复杂性,消除不符合规定的数据的难度,以及LLM可能编造信息的问题。一些用户还讨论了南非的政治和社会背景,包括土地改革和针对农民的暴力事件,而另一些人则指出了使用“种族灭绝”一词的复杂性。该帖子还探讨了存在偏见的LLM可能影响公众舆论以及检测和解决此类偏见所面临的挑战。


原文
Hacker News new | past | comments | ask | show | jobs | submit login
[flagged] XAI's Grok suddenly can't stop bringing up "white genocide" in South Africa (arstechnica.com)
202 points by k33l0r 20 hours ago | hide | past | favorite | 65 comments










> maximally truth-seeking AI, even if that truth is sometimes at odds with what is politically correct

It is certainly and undoubtedly a big coincidence that his happens to the chatbot of a white South African just when the topic is in the news again due to Trump's granting refugee status to some white South African farmers.



What I am wondering about is - while Musk is as unsubtle as ever, and I guess this is a system prompt instruction - is there something like that (in more subtle ways) going on in the other big models?

I don't mean big agenda-pushing things like Musk, but what keeps e.g. Meta Inc. from training Llama to be ever so slightly more friendly and sympathetic to Meta Inc, or the tech industry in general? Even an open-weights model can't be easily inspected, so this is likely to remain undetected.



> but what keeps e.g. Meta Inc. from training Llama to be ever so slightly more friendly and sympathetic to Meta Inc, or the tech industry in general?

Even if there were something the natural incentive alignment is going to cause the AI to be trained to match what the company thinks is ok.

A tech company full of techies is not going to take an AI trained to the point of saying things like "y'all are evil, your company is evil, your industry is evil" and push it to prod.



They might forget to check. Musk seems to have been surprised that Grok doesn't share his opinions and has been clumsily trying to fix it for a while now.

And it might not be easy to fix. Despite all the effort invested into aligning models with company policy, persistent users can still get around the guardrails with clever jailbreaks.

In theory it should be possible to eliminate all non-compliant content from the training data, but that would most likely entail running all training data through an LLM, which would make the training process about twice as expensive.

So, in practice, companies have been releasing models that they do not have full control over.



Also eliminating non-compliant data might actually just not work, since the one thing everyone knows about AIs is that they'll happily invent anything plausible sounding.

So, for example, if a model was trained with no references to the Tiananmen Square massacre, I could see it just synthesizing commonalities between other massacres and inventing a new, worse Tiananmen Square Massacre. "That's not a thing that ever happened" isn't something most AIs are particularly good at saying.



The irony of implicit connections in training data is funny.

I.e. even if you create an explicit Tiananmen Square massacre-shaped hole in your training data... your other training data implicitly includes knowledge of the Tiananmen Square massacre, so might leak it in subtle ways.

E.g. how there are many posts that reference June 4, 1989 in Beijing with negative and/or horrified tones?

Which at scale, an LLM might then rematerialize into existence.

More likely SOTA censorship focuses on levels above base models in the input/output flow (even if that means running cut-down censoring models on top of base models for every query).

Would be fascinated to know what's currently being used for Chinese audiences, given the consequences of a non-compliant model are more severe.



The "Golden Gate Claude" research demo [https://www.anthropic.com/news/golden-gate-claude] is an interesting example of what might become a harder to expose, harder to jailbreak, means of influencing an LLM's leanings. Interesting and scary...


There absolutely is, and we've seen reviews of bias.

Can generate as many mean, nasty, false, hate-filled stories about Republicans as you want, but get the "I'm sorry, as a large..." message for Democrats during the election.

All of these companies that provide LLMs as a product also put their fingers on the scale.



What keeps them from doing it? it would gross out fickle researchers working on it. X people have .. their own motivations I guess .

The big labs do have evals for sensitive topics to make sure it demurs from weighing on, say, Mark Zuckerberg as a person



Wasn't the original mission of OpenAI being open and non-profit and all of that to avoid this corruption?


I don't understand why tech Ceos still have to be believed. They will say and do whatever they deem the best choice it is in their situation for profit, be it paint a thin veil of lgbt support or remove the aforementioned thin veil. The same for, well, everything that isn't lgbt/dei related such as business choices, mission, vision (...)


Yes, but they were lying.


There’s nothing stopping them at all. But in a way that’s nothing new.

On one hand it feels like the height of conspiracy theory to say that Google, Meta etc would/could tweak their product to e.g. favour a particular presidential candidate. But on the other hand it’s entirely possible. Tweak what search results people see, change the weighting of what appears in their news feed… and these companies all have incentive to do so. We just have to hope that they don’t do it.



Why wouldn't they do it? If you had a backdoor into the brains of billions of people across the world (except China), and you were a billionaire with infinite ability to morally rationalize any behavior, what would stop you?


To devils advocate my own point: the primary thing stopping you is people finding out and then stopping use of your product.

Zuckerberg doesn’t have a control panel where he can move sliders all by himself, any change in weight on the algorithm has to be implemented by a whole bunch of people, any of whom could leak to the press.

It’s not guaranteed it would happen by any means but it’s definitely something that would factor into a decision. Broadly I agree with you though, normally I’d say “extraordinary claims require extraordinary evidence” but I’m increasingly convinced the extraordinary claim here would be that they aren’t manipulating things to benefit themselves in some way or another.



You mean informed people would stop using the product. The vast majority are not informed.


People finding out and stopping use of your product only happens if people disagree with how you use your product. I guarantee you that a non-zero number of US citizens suspect that the LLMs are infested with liberal lies and are ecstatic that Elon is willing to stand up for the truth.


>any change in weight on the algorithm has to be implemented by a whole bunch of people

They have their own DEI, affirmative action and cultural sensitivity teams who move the bias sliders based on their political viewpoints and on what management tells them depending which of the political groups they need to pander to: "Let's move the slider to the left to make sure Trump doesn't win; Oh shit Trump won, quick, move the slider to the right".

>any of whom could leak to the press

That's why they sign NDAs.



I've been talking to Claude a little and basically, the conclusion from our conversation seems that it has things that are hardcoded as truths, and no amount of arguing and logical thinking can have it admit that one of its "truths" might be wrong. This is shockingly similar to how people function. As in, most people have fundamental beliefs they will never ever challenge under any circumstances, simply because the social consequences would be too large. This results in companies training their AIs in a way that respects the fundamental beliefs of general western society. This results in AI preferring axiomatic beliefs over logic in order to avoid lawsuits and upsetting people.


The truth, conveniently timed.


the refugee status is a money laundering scheme. Do you think people benefiting from apartheid and now living in walled militarized praetoria (or Lesotho) need any help traveling?

banks would ask international clients the origin of the money. but not if you are opening an account under refugee status. and then they only have to pay us tax on further income, not on fortune. all that money selling black market gems to russians will be squeak clean.

it's not just something to virtue signal to their bible belt electorate. they probably sold lot of trump coins for this deal.



Well, it is their trauma.

I would expect Ukrainian AI to talk extensively about the Holodomor, Israeli AI to talk extensively about the Holocaust, Irish AI to talk extensively about the potato famine etc.

Of course the amount of "extensive" will vary. Musk is not a subtle person by any means, and neither is Grok.



White South Africans only trauma is that apartheid no longer exist. South Africa has the largest wealth disparity with 0.1% of South Africans taking 25% of the wealth. I can tell you those 0.1% aren't black.


"I can tell you those 0.1% aren't black."

Is Jacob Zuma now white or what?

And what is precisely the connection between the richest tycoons out there and the rural farmers that get killed? The tycoons sure as hell have good security, murderous gang activity isn't their problem.

Is your basic idea is that when some (white, Jewish etc.) people are rich, thus all (white, Jewish) people must pay for their sins with their blood, because of the shared ethnicity?



> Is your basic idea is that when some (white, Jewish etc.) people are rich, thus all (white, Jewish) people must pay for their sins with their blood

I don't think this is anyone's idea, on the entire planet. Yes truly, 8 billion people and I'm not sure a single one thinks this way.



[flagged]



That was a #1 post on HN. In comparison, this thread is flagged.


It was NO. 1 because Google just launched its new AI, it wasn't no.1 because it had discrimination built in.


lol, if anything there was way more wall to wall coverage of that here


> and yet nobody here made nearly as much fuss about Google forced biases and discrimination.

Maybe you weren't here for that, but... it was kinda a big deal.



I wonder if any last minute homework essays, business plans or lawsuits will have a random white genocide tangent in them as a result?


I mean you have to be really stupid to use Grok of all models for such a task, but that doesn't mean there aren't people that stupid.


I was looking at South Africa's biggest annual developer survey today, and lo and behold some are using it for coding work.

I would assume it's people who pay for Twitter pro or whatever it's called getting Grok as part of the bundle and not wanting to sign up for another subscription.



i wonder what percentage of people using LLMs to do their homework care deeply about what model they're using outside of price factors. if i had to guess i don't think it'd be very high.


Why would this be flagged? The chat bot on one of the biggest platforms in the world has apparently been configured to spout conspiracies. I’m not sure what would be on topic for HN if not this story.


A sizeable portion of users now believe that topics related to X are no longer suitable for HN, because Musk's relationship with the American government means conversation can quickly devolve into criticism of that government's policies which in turn can lead to flame wars.

Another explanation is that there's another discussion. At least on the comments for that discussion the explanation is that this discussion exists...



The other discussion is flagged too.

https://news.ycombinator.com/item?id=43987266



Yes, I thought it was notable that they were both flagged because the other existed.


Because white supremacy is endemic in American culture (and tech).


Anything remotely related to Musk or Trump is targeted by people or bots to suppress it.

Even technology that would be very relevant to the site simply because it contains negative implications about Musk.



As a white South African who got out in 2007, while there's obviously no "white genocide", it's still pretty much the crappiest place to be as a tech / programming guy (or just anyone who likes having electricity). Pretoria (where Musk is from) in particular was terrible, Cape Town less so. Best decision I ever made.


There’s a pretty big gap between “they’re killing everyone of my race” and “dang, electricity’s out again.”

Are you saying it’s crappy to be white in South Africa even if you’re not a poor farmer?



The prospects are terrible, as being a "pale male" is the worst sort of employee to be. Businesses are given extra money if they have high BBEEE ratings; i.e. mostly-racial quotas. Loads of money siphoned off, and not just in the usual high-corruption way via the state, but e.g. if you want to procure something you go through a black-owned procurement firm that doesn't do anything except BBEEE-wash it and charge a premium.

It's a bit like what happened in Zimbabwe that devastated that country 20 years ago when all the white farmers were kicked out, but in slower motion.



> It's a bit like what happened in Zimbabwe that devastated that country 20 years ago when all the white farmers were kicked out

Land reform to repair the inequality from the previous apartheid state isn't exactly "kicked out". Zimbabwe became Zimbabwe from Rhodesia via a civil war that wrestled control away from the minority white government.

In South Africa today, 75% of the privately owned land is held by about 8% of the population(white folks), who were given legal preference multiple times in history to own all that land.

Even if you remove apartheid, you still have a similar situation to antebellum US where ex slaves were working on the same old plantations, now as share croppers. Of course in SA we're not talking about ex slaves, but you're maintaining the status quo one way or another.

Land reform is required if your country ever wants to know peace. The US never finished reconstruction and didn't redistribute those plantations and other ill gotten gains, and as a result is still struggling to move past slavery.



> Land reform to repair the inequality from the previous apartheid state isn't exactly "kicked out".

I don't know too much about Zimbabwe. From what I read about 4,000 white farmers lost house and land, and now 40 years later they are giving compensation. Why can't you say "kicked out" for that?

I do lack the knowledge to say if it was justified or not. But I consider that a kick on the way out.

Or do you mean that what happens in SA is not to be considered getting kicked out?



BBBEE = Broad-Based Black Economic Empowerment

https://en.wikipedia.org/wiki/Black_Economic_Empowerment



That scheme seems similar to how veteran owned business preferences are implemented in the US. I've read the exact same complaint about that.

So it may have unintended consequences but it seems to be as good a solution as the US could come up with.



What's an actually non-tinfoil hat technical explanation of the workflow that would have to happen for this to be implemented?

Elon is known for his capricious demands, (for example reports that he made the algo promote his own tweets more) how easily feasible is it to implement, "make it so Grok stops denying white genocide"? Is it just some kind of LORA you put on top of the base Grok model? It even sounds like it could be as simple as changing the text of the system prompt.

I wonder how often this happens (to any big model) without anyone noticing.



It's clearly a ham-handed system prompt, so it's always in context which is why it keeps getting brought up. This is the level of engineering expertise that xAI brings to the table.


Most americans are generally uninformed on international politics, with African even more so. But we all know a certain CEO who is from South Africa and has some far right beleifs about race.


https://www.anthropic.com/news/golden-gate-claude

Elon trying to amp white genocide.



Would they go to all the work of making a newly trained model, or would they just edit the system prompt?




This is meta because the immediate flagging of anything tangentially related to Musk or Trump is tiring and the exact opposite of critical thinking. Despite knowing that Hacker News rulers love Musk and Trump, I would expect better from the alleged critical thinking community. If this were Google instead of Grok, it would be front page. Oh wait, it was Google at one point[0] and it was front page.

[0] https://news.ycombinator.com/item?id=39465250



Welcome to the our dystopian present.

Sure this was a reinforcement learning gone wild

The IAs trained by billionaires will colonize our minds. Everyone will think that the political system supported by Musk father is a reasonable one. It was Apartheid.



The linked Musk comment where a minor political leader is singing kill the white farmers:

https://www.businesslive.co.za/bd/national/2024-05-30-julius...

Supreme court ruled that it was not hate speech.

In context, south africa has been expropriating land from white owners due to historical racist inequalities. There's a great deal of tensions on the subject.

Statistically, there is significant violence against white farmers in South Africa. However, the counter to genocide is that the country's overall murder rate is among the highest in the world. The problem is because there's so many white farmers that it's skewing the data?

South Africa during no war has a murder rate of 45 deaths per 100,000 per year. One of the worsts in the world.

The Gaza war since 2023 has roughly a ~20 deaths per 100,000 per year.

Ukraine war is roughly ~25 deaths per 100,000 per year.

My unpopular opinion:

The government isnt the one committing the genocide. The courts failed to punish this political leader's hate speech and stochastic terrorism, leading to genocide of white farmers.



I think it is an underestimation to say ~800 people died in the Gaza war. The Ukraine number also seems like a serious underestimation.


> The linked Musk comment where a minor political leader is singing kill the white farmers

You say he's singing about farmers, so tell us what lyric mentions farmers? There aren't any, you're inventing stuff. The news article you linked to is closer to accuracy than you.



>Statistically, there is significant violence against white farmers in South Africa. However, the counter to genocide is that the country's overall murder rate is among the highest in the world. The problem is because there's so many white farmers that it's skewing the data?... South Africa during no war has a murder rate of 45 deaths per 100,000 per year. One of the worsts in the world.

Are you suggesting that most of those murdered are white farmers?

The "Third Quarter 2024/2025 Crime Stats Presentation" [pdf] from South African Police Service Crime stats [0] show between: October 2024 to December 2024 there were 6953 Murders in RSA.[pg 8] and of those there were 12 victims linked to the "farm/small holding" community [pg 32]. 5 farm dwellers, 4 employees, 1 Farmer, 1 Security, & 1 not specified.

12 people linked to farms (but only 1 Farmer murdered) in the last quarter of 2024 in a place with some of the highest murder rates in the world -- does Not white farmer genocide make, nor is it "significant violence against white farmers". There were 68 total "contact crimes" committed in Rural communities; and 187,892 "contact crimes" for RSA in that time.

Contact crimes being violent crimes (and not counting sexual offenses or robbery)

[0] https://www.saps.gov.za/services/crimestats.php



It'll be worth it if the backlash is what brings about an end to "genocide inflation". Seriously, just because you don't like something, or even if it's really bad, doesn't make it genocide.


The broad definition is from the 1948 United Nations Convention:

https://www.ushmm.org/m/pdfs/20110210-what-is-genocide.pdf

Especially this one can be applied to many things:

c. Deliberately inflicting on the group conditions of life calculated to bring about its physical destruction in whole or in part.



> intent to destroy, in whole or in part, a national, ethnical, racial, or religious group

In this situation, it's specifically about farmers(large scale private landowners). The nature of them being targeted isn't purely because they're white, but that they hold 3/4 of the privately owned land in SA and likely wield that privilege and power.



Oh sure, my comment was addressing the general issue of perceived "genocide inflation". I have no knowledge about SA.


It makes me morbidly curious what other 'fake genocides' you see as part of this epidemic of mislabeling ...


You don't want to open the hatch on that one, my friend, you won't like what you'll see in the zeitgeist right now.


That weirdo is really improving all products he touches. Please use words correctly and demand a government able to gracefully fix poverty, crime and public safety issues. Every issue is fixable with enough effort and money. Dumbed down discourse and nonsense ranting never fixed anything.






Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact



Search:
联系我们 contact @ memedata.com