It might also be a bit of “we’re trying to redefine AGI to make sure we hit the MS bonus targets”
Every few weeks in the law school subreddit, someone asks what the point of going to law school even is anymore, since AI is going to replace lawyers soon. I’m going to start posting this picture in response to that claim.
The reason for this may be partially to get money from Internet Archive, but part of the reason is also AI data scraping
Yet another demonstration that AI doesn’t actually “understand” anything.
But I’m sure it will be discovering brand new types of science and physics any day now…
When metrics become a target…
Models for interpolation don’t work for extrapolation. The only surprise is how much tech press lets companies get away with pretending otherwise.
I’m concerned that we’re very close to most people seeing that and, rather than saying “ha, look how dumb and useless AI is”, saying “AI said it so I guess it must be true”. There’s no bottom to human stupidity and far too many people seem eager to let AI replace their brain entirely.
(excerpt) As companies like Amazon and Microsoft lay off workers and embrace A.I. coding tools, computer science graduates say they’re struggling to land tech jobs.
Gee. Whodda thunk it?
Computer Science Grads Struggle to Find Jobs in the A.I. Age - The New York Times
From someone I showed this to:
What happens if you tell it that is wrong?
(Not from the “AI Overview”, that didn’t trigger for me this time, but this was the “AI Mode” result)
You know, I don’t even want to spend all day working for these losers. If there were things like universal health care and universal basic income then I wouldn’t care how many crappy jobs they eliminate. So why don’t all these AI guys ever support those?
If I remember correctly, the MS targets had their own definition of “AGI” which was merely yearly revenue (which we could generously interpret to mean that if a certain number/breadth of jobs had been replaced, it could functionally be considered “AGI”). Of course, given the weak revenue and growth and narrow market for ChatGPT (mostly students cheating), OpenAI doesn’t remotely have a path to hit that target either, so…
Alternatively you could just post any of the (many, many) news stories about lawyers getting in trouble precisely because they used AI and it generated legal nonsense. (I know the credulous keep saying that AI will totally get better and all the problems will magically melt away, but it’s very clearly not happening and never will.)