Any one of those could have been used to direct the agent to behave in a certain way, or to create a specific type of post.
My point is that we really don’t know what happened here. It is possible that this is yet another case of accountability washing by claiming that “AI” did something, when it was actually a human.
However, it would be really interesting to set up an openclaw agent referencing everything that you mentioned for conflict resolution! That sounds like it would actually be a super power.