It took me a few minutes to figure out what your Picassoesque reindeer was. I thought it was some kind of deformed moose.
I don't mean to be negative though .... the content and writing is great ... it's just the image of the reindeer was distracting for a minute before I could look through the rest of the comic.
Back in my day, we didn't have no dang hoverboards, and hoverboards were a thing of the future from 2015 when Marty McFly would fly across the silver screen on his one true hoverboard. I spent the best years of my life working on anti-gravity technology with the hope that I could one day too fly like Marty, on a real hoverboard, but now I just sit in my own graveyard of failed dreams while the kids are zoomin' around on their wheely-scooters and calling them hoverboard. Thanks Obama!
Well, it's not a surprise that the definition of "AI" is not based on how it is represented in fiction. It shouldn't.
But the definition of AI is still oddly large and include a lot of things that probably shouldn't be part of it.
On the other hand, when people talk about "AI", it's almost always about machine learning, aimed at NLP or vision tasks, which is also inaccurate as AI can do much more than that.
ChatGPT behaves very much like the AI we've seen in fiction. You can use it pretty much exactly like the crew of the Starship Enterprise uses the ship's computer.
In personally trying to use ChatGPT 4 for a job task (programming), I would disagree strongly with this sentiment. I have yet to find a task where it doesn't partially fail due to no notion of the concepts underlying the topic.
In an example, I asked it to write an implementation of reading from a well known file type as a class. It had many correct ideas for certain operations (compiled from other sources of course), but failed with the basic concept of class instantiation. It was calling class methods in the constructor, which is just not allowed in the language being used. I went through several iterations with it to avail before just giving up on it.
In "normal" language tasks, it seems to be quirky, but passable. But if you give it a highly technical task where nuance and conceptual knowledge are needed? I have yet to see that work in any reliable capacity.
I use it for programming a lot too. You have to explain everything to it like you would a brand new engineer, and then it is often wrong with certain parts like you said. But if you know enough about coding to figure out where it's wrong, and just write those parts yourself, it can still be a huge time saver.