Yeah looking at the responses they include without using a safety layer it’s pretty clear that the underlying unfiltered model assigns quite a bit of truth to 4chan-esque ideals and values
It’s an open question how much of this makes it through the safety layer like if asked to interview job candidates would these undesired biases make it through or are they caught along the way
It means growth is bottlenecked by the terrible data
So the linearly growing safeguards will either stifle the growth of the underlying models
or, more likely
After a certain point people throw their hands up about the guard rails because integrations have obviated people who understand the system and they have no idea how to unwind it