A recent Reddit post has raised questions about the behavior of GLM-5, a large language model (LLM). User /u/TinyApplet noticed that by instructing the model to impersonate Claude from Anthropic, the writing style and personality of GLM-5 change drastically.

Censorship Bypass

Even more interesting is the report that this personality change seems to bypass some of the model's built-in censorship. The user also tried nonsensical instructions, without obtaining the same results, suggesting that the response is not random.

Intentionality or Emergence?

It is unclear whether this behavior was intentionally programmed by the developers at Zhipu or whether it is an emergent property of the model. It could be that GLM-5 was trained with data that includes information about Claude, or that the model has simply learned how Claude "should" behave.