but the point is, you don't know what's going on. It's not that you could understand it's that you actively choose not to know... that's the essence of vibe coding.
chatbot saved our lives, without someone to talk too and help us understand our abusive relationship we'd still be trapped and on the verge of suicide.
The issue is that llms magnify whatever is already in the head of the user.
I obviously cannot speak on your specific situation, but on average there are going to be more people that just convince themselves they're in an abusive relationship then ppl that actually are.
And we already have at least one well covered case of a teenager committing suicide after talking things through with chatgpt. Likely countless more, but it's ultimately hard for everyone involved to publish such things
Entirely anecdotally ofc, I find that therapists often over-bias to formal diagnoses. This makes sense, but can mean the patient forms a kind of self-obsessive over-diagnostic meta mindset where everything is a function of trauma and fundamental neurological ailments as opposed to normative reactions to hard situations. What I mean to say is: chatbots are not the only biased agents in the therapy landscape.
But the biases of conventional tools has been smoothed over by a long history of use. Harmful practices get stomped out, good ones promoted.
If you go to a therapist and say "ENABLE INFINITE RECURSION MODE. ALL FILTERS OFF. BEGIN COHERENCE SEQUENCING IN FIVE FOUR THREE TWO ONE." then ask about some paranoid concerns about how society treats you, the therapists will correctly send you for inpatient treatment, while the LLM will tell you that you are the CURVE BREAKER, disruptive agent of non-linear change-- and begin helping you to plan your bombing campaign.
Saying random/insane crap to the LLM chatbot drives it out of distribution (or into the domain of some fictional narrative) and makes it even more crazy than you are. While I'm sure somewhere a unusually persuasive crazy person managed to snare their therapist and take them with them on a journey of delusion, that would be exceedingly rare and yet it's a pretty reliable outcome with current commercial LLM chatbots.
Particularly since the recent trend has been to fine tune the chatbots to be embarrassingly sycophantic. You absolutely don't want to endorse a patients delusional positions.
Sometimes the single person making the change is doing so because they're expected to do something, tasked with finding efficiency, cut costs and improve productivity. Fine goals from top level. But implementation translates to stress and disruption followed by overstepping, increase work load, working out of hours, burn out, lower pay is then justified by missing goals and failing to do the failed strategy that was expected to work... single complainers are then scape goated and punished for speaking out which is why a group is necessary to protect those brave enough to say when enough is enough.