> But those using this as an argument to ban AI
Are people arguing that, though? The introduction to the article makes the perspective quite clear:
> In tweaking its chatbot to appeal to more people, OpenAI made it riskier for some of them. Now the company has made its chatbot safer. Will that undermine its quest for growth?
This isn't an argument to ban AI. It's questioning the danger of allowing AI companies to do whatever they want to grow the use of their product. To go back to your previous examples, warning labels on cigarette packets help to reduce the number of people killed by smoking. Why shouldn't AI companies be subject to regulations to reduce the danger they pose?