That old feature uses Whisper to transcribe your voice to text, and then feeds the text into the GPT which generates a text response, and then some other model synthesizes audio from that text.
This new feature feeds your voice directly into the GPT and audio out of it. It’s amazing because now ChatGPT can truly communicate with you via audio instead of talking through transcripts.
New models should be able to understand and use tone, volume, and subtle cues when communicating.
I suppose to an end user it is just “version 2” but progress will become more apparent as the natural conversation abilities evolve.