I think that's part of it, but then the user perceives "personality changes" when the model changes due to differences in the model. Now they have lost their relationship because of the model change.
Anyone attempting to build a movement might find it interesting how Pumping Station One in Chicago is governed. It's a maker space but run by people who care (at least from my experience when I was a member back in 2015/2016). The process for electing leadership and holding members accountable was very democratic and fair, from my experience. They open-source as much as they can about how they organize:
Microsoft Teams was bad, so they rebuilt it and somehow made it worse. Then they decided to do the same with other apps, like Notepad. I switched to Ubuntu on my computer this week. Linux administration is not something I want to spend time on, but LLMs are able to help me debug why my password manager can't talk to my browser and write shell scripts to fix it... I'm able to focus on work and be done with the Microslop.
There are a lot of Android devices that look temping until one discovers how out-of-date the firmware is.
With no option to install your own, of course. Boot loaders should be exclusively for running the manufacturer's lone security update from 5 years ago.
I do this with Chrome recording and Playwright. What I need is an AI agent to meander through my product as if it were the target user and test/break things so I can pass that to my LLM to fix. Does anyone have that?
Different from our core use case, but our agents can do open-ended exploration as well. You could prompt something like "navigate to this app as a new user and try common. flows" with structured outputs for findings. Session recording will show what happened. Not sure if it fully solves your problem - but happy to explore this together if you want to try it.
Many of these announcements are bluffs as many users here have pointed out. But real LLM-driven layoffs do happen, and from what I have anecdotally seen, they follow a pattern: leadership assumes the new LLM service will make human workers redundant. They then make cuts before the evidence is in. What this means is that today, there are many LLM service deployments that replaced humans while their actual impact remains a mystery. Though it won't be a mystery to leadership forever.
One example client that shouldn't dox me: Odom Corporation, a beverage distributor. They purchased an LLM-driven purchasing solution and immediately laid off their entire purchasing team, save for a few members who exist on the periphery. A follow-up with them showed that the system was ordering summer beverages coming into the winter (among many other bad purchasing decisions) and causing a dramatic increase in unsold inventory. Since they believe that LLMs will exponentially improve, they're dismissing it as a one-off because this year's models "will be so much better". We attempted to advise differently, but stakeholders got extremely emotional at even small suggestions that there was a fundamental problem. Good luck to them.
Yeah. I keep running into LLMs in customer service functions where I would previously have been talking to a human, and in literally every case they're beyond worthless and I end up talking to a human anyway.
I think the execs are either using it as an excuse to reduce opex to boost share prices or they're actually buying into this delusion that the productivity improvements are right around the corner. Though I don't really buy into that second option since any reasonably intelligent person would wait to see concrete evidence of said improvements before crippling your company based on some hope.
I also think some of the companies that operate in the AI space are using the layoffs as a form of marketing to prove the capabilities of their tech (while also using it as an excuse to cut costs).
Anyways, I work at one of the major players in the space and the amount of AI code slop I see on a daily basis is absurd. My prediction is that within two years most younger SWEs will only have a high-level understanding of their code. I already see it happening.
reply