Just out of curiosity. I have no moral stance on it, if a tool works for you I’m definitely not judging anyone for using it. Do whatever you can to get your work done!
Just out of curiosity. I have no moral stance on it, if a tool works for you I’m definitely not judging anyone for using it. Do whatever you can to get your work done!
A lot of people are going to get fucked if they are…
It’s using the “startup method” where they gave away a good service for free, but they already cut back on resources when it got popular. So what you read about it being able to do six months ago, it can’t do today.
Eventually they’ll introduce a paid version that might be able to do what the free one did.
But if you’re just blindly trusting it, you might have months of low quality work and haven’t noticed.
Like the lawyers recently finding out it would just make up caselaw and reference cases. We’re going to see that happen more and more as resources are cut back.
Huh? They already introduced the paid version half a year ago, and that was the one being responsible for the buzz all along. The free version was mediocre to begin with and has not gotten better.
When people complain that ChatGPT doesn’t comply to their expectations it’s usually a confusion between these two.
It’s been notorious for doing that from the very beginning though
Anyone blindly trusting it is a grade A moron, and would’ve just found another way to fuck up whatever they were working on if ChatGPT didn’t exist.
ChatGPT is a tool, if someone doesn’t know what they’re doing with it then they are gonna break stuff, not ChatGPT.
This is exactly like people who defend Tesla by saying it’s your fault if you believed their claims about what a Tesla can do…
Which isn’t a surprise, there’s a huge overlap between being gullible to believe either companies claims, and some people will vend over backwards to defend thos companies because of sink cost fallacy
I don’t know what OpenAI even claims that ChatGPT can do, but if you trust marketing from any company then you’re gonna get burnt.
I’m not defending the company in any way, more just defending that in general LLMs can be useful tools, but people need to make educated decisions and take a bit of responsibility.
That may have been their plan, but Meta fucked them from behind and released LLama which now runs on local machines, up to 30B parameter size and by end of the year will run at better than GPt3.5 ability on an iphone.
Local llms, like airoboros, WizardLm, Stable Vicuña or Stable Coder are real alternatives in many domains.
Removed by mod