Hacker News
Beijing is enforcing tough rules to ensure chatbots don’t misbehave
stevenjgarner
|next
[-]
meyum33
|root
|parent
|next
[-]
Zetaphor
|root
|parent
|next
|previous
[-]
stevenjgarner
|root
|parent
|next
[-]
sokoloff
|root
|parent
|next
|previous
[-]
Architecture and training data both matter.
AlotOfReading
|root
|parent
[-]
It doesn't seem impossible that models might also be able to learn reasoning beyond the limits of their training set.
Retric
|root
|parent
|next
[-]
When you only celebrate success simply coming up with more ideas makes things look better, but when you look at the full body of work you find logic based on incorrect assumptions results in nonsense.
throwuxiytayq
|root
|parent
|previous
[-]
Similarly, the leading models seem perfectly secure at first glance, but when you dig in they’re susceptible to all kinds of prompt-based attacks, and the tail end seems quite daunting. They’ll tell you how to build the bomby thingy if you ask the right question, despite all the work that goes into prohibiting that. Let’s not even get into the topic of model uncensorship/abliteration and trying to block that.
cherioo
|root
|parent
|next
|previous
[-]
throwuxiytayq
|root
|parent
[-]
meltyness
|root
|parent
|next
[-]
I think current censorship capabilities can be surmounted with just the classic techniques; write a song that... x is y and y is z... express in base64, though stuff like, what gemmascope maybe can still find whole segments of activation?
It seems like a lot of energy to only make a system worse.
sho_hn
|root
|parent
|next
|previous
[-]
Haven't some of them already? I seem to recall Grok being censored to follow several US gov-preferred viewpoints.
bilbo0s
|root
|parent
|previous
[-]
It's the arts, culture, politics and philosophies being kneecapped in the embeddings. Not really the physics, chemistry, and math.
I could see them actually getting more of what they want: which is Chinese people using these models to research hard sciences. All without having to carry the cost of "deadbeats" researching, say, the use of the cello in classical music. Because all of those prompts carry an energy cost.
I don't know? I'm just thinking the people in charge over there probably don't want to shoulder the cost of a billion people looking into Fauré for example. And this course of action kind of delivers to them added benefits of that nature.
SilverElfin
|next
|previous
[-]
bgwalter
|next
|previous
[-]
More censorship and alignment will have the positive side effect that Western elites get jealous and also want to lock down chatbots. Which will then get so bad that no one is going to use them (great!).
The current propaganda production is amazing. Half of Musk's retweets seem Grok generated tweets under different account names. Since most of responses to Musk are bots, too, it is hard to know what the public thinks of it.
Imustaskforhelp
|previous
[-]
If the govt. formally anounces it, perhaps, I believe that they must have already taken appropriate action against it.
Personally I believe that we are gonna see distills of large language models perhaps even with open weights Euro/American models filtering.
I do feel like everybody knows seperation of concerns where nobody really asks about china to chinese models but I am a bit worried as recently I had just created if AI models can still push a chinese narrative in lets say if someone is creating another nation's related website or anything similar. I don't think that there would be that big of a deal about it and I will still use chinese models but an article like this definitely reduces china's influence overall
America and Europe, please create open source / open weights models without censorship (like the gpt model) as a major concern. You already have intelligence like gemini flash so just open source something similar which can beat kimi/deepseek/glm
Edit: Although thinking about it, I feel like the largest impact wouldn't be us outsiders but rather the people in china because they had access to chinese models but there would be very strict controls on even open weights model from america etc. there so if chinese models have propaganda it would most likely try to convince the average chinese with propagandization perhaps and I don't want to put a conspiracy hat on but if we do, I think that the chinese credit score can take a look at if people who are suspicious of the CCP ask it to chatbots on chinese chatbots.
KlayLay
|root
|parent
[-]
A technology created by a certain set of people will naturally come to reflect the views of said people, even in areas where people act like it's neutral (e.g., cameras that are biased towards people with lighter skin). This is the case for all AI models—Chinese, American, European, etc.—so I wouldn't dub one that censors information they don't like as propaganda just because we like it, since we naturally have our own version of that.
The actual chatbots, themselves, seem to be relatively useful.
Workaccount2
|root
|parent
[-]
It might not feel like that on the ground, the leash has been getting looser, but the leash is still 100% there.
Don't make the childish mistake of thinking China is just USA 2.0
Imustaskforhelp
|root
|parent
[-]
So like even now although I can trust chinese models, who knows for how long their private discussions have been happening and for how long chinese govt has been using that leash privately and for chatbots like glm 4.7 and similar.
I am not sure why china would actively come out and say they are enforcing tough rules tho, doesn't make much sense for a country who loves being private.