This is the reason I balk at personifiing these things with human terms. It sounds cool but it is both inaccurate and misleading especially in the hands of the media and the general public.
The dilemma is, ChatGPT can write better reports than most graduate school students in my country. For what these problematic vast majority of students do is remember, not analyze.
Specifically for this context, students are not trained to analyze what they are asked (input query for ChatGPT). When I ask a unique question in their assignment, they can’t even form a response. They just write a generic text that doesn’t try to answer my question.
They seem to copy and paste what’s in their brain. And when it comes to copy and pasting, i.e. mimicking what people do, ChatGPT is the champion in some sense. Hell, OpenAI even tuned it to generate balanced stance, and that’s also what students can’t do.
Finally, 90% of the population perform actually worse than these graduate students.
It is sad, but most people seem to go to school for certification not learning. Use to grade when in grad school… the lazy sloppy work was nuts. Working for a company… the terrible writing some people do even with advanced degrees.
“…AI” concerns me. I use quotation marks there because what is often referred to as AI today is not whatsoever what the term once described.
Lost me right there. Not only was and is this AI, but the term gets narrower over time, not broader. If you want to go by “what the term once described,” you have to include computer vision, text to speech, optical character recognition, behavior trees for video game enemies, etc etc etc.
When I see people complain about calling LLMs “AI,” I think the only definition that would satisfy them is “things computers can do that we aren’t used to yet.”
removed by mod
removed by mod
Our one rule here is to be nice. Calling someone a dumbass is not nice. Be better.
I think it’s less of an issue of LLMs being drunk and more that ostensibly sober people put them behind the wheel totally aware of how drunk they are while telling everyone that they’re stone cold sober.
gasp you mean, industry is lying to investors about a new technology to get more investment and creating a false narrative for the public to undermine criticism? Who could have seen this coming!
I’m a scientist entering the industry and can’t agree more. Too many lies. There are handful of companies that do deliver, but, generally speaking, many businesses seem to bet on the naivety of investors. Some even do it unintentionally because the bar is simply too low.
But here’s the thing. I noticed in my life that there are naive people following weak narratives no matter what. These people can’t doubt your arguments. And if you do business, these people are apparently the most solid support.
Edit: it’s perhaps also true that this majority of investors forces companies to lie by investing almost solely on LLMs
I think there is just too much faith from the current crop of investors in tech start ups, many got hurt in the past by not investing in things after not getting good answers to “ok but how does this make money” or “can this actually do what you’re claiming”.
And larger more established companies like Google and Amazon are happy to feed the hype for a lot of these trends, particularly when all the new start ups are going to be buying stuff from them, so even if the start ups fail because they can’t make money or don’t do what they claim, the big companies still made money selling them server space, computing time, or huge amounts of data. I think investors who hold stakes in the big companies also lean in to the hype for this reason.
Everyone has a pretty good incentive to lean in to the hype, so they do.
Until like 3 months ago, I felt the ChatGPT revolution was going on. Every 10-year plan my colleagues in AI research was actually completed in a few weeks(!) by a completely different research team at the opposite side of the planet. The hype was so high that every expert had same plans resulting in surreal competitions.
After that, the LLM businesses entered the b2b space, with all potential customers asking ChatGPT to search information in their pile of documents. That was the next big thing.
We haven’t heard back from the pile of garbage so far…
hustles to find pearls to clutch
I liked this article, and I think a lot of the commenters here are missing that the general public is treating LLMs as AGI. I have a whole 5-10 minutes I spend on why this is when I present about LLMs.
“The I in LLM stands for Intelligence” is a joke I read (and include in my presentation to hammer the point home). Laymen have no idea what AI or LLMs are, but they expect it to work similarly to human intelligence, since that’s the only model they know, and are surprised to learn it doesn’t work that way.
Edit: Forgot what I came to the comments to post, before I read everyone else’s complaints about this, lol.
A small correction: the Air Canada example wasn’t an LLM, it was just an old “dumb” chatbot that was likely sharing outdated policies.