There's an interesting aspect of this issue that I think the post summary really dismisses. Photos coming from phones these days sort of are AI, and in an annoyingly pervasive way.
I've actually gone back from using my phone to using a proper camera again over the last year or so because I'm getting so irritated by the amount of ML-based post-processing my phone does. It results in a lot of photos looking bad, and there's no easy way to bypass it besides setting the phone to save raw which sort of defeats the point of using the phone in a lot of ways (ability to go from taking the photo to posting on the device). A really common situation for me is when I take a photo with my phone that is blurry because of bad focus/shake/low light/some combination. The phone does really aggressive ML "sharpening" of the image that makes it look extremely artificial and, frankly, a lot worse than if the postprocessing had been omitted. I've had sets of photos I took totally ruined by this kind of "helpfulness."
It's a tricky issue, there absolutely are benefits to cameras using the best technology available to create the best photograph available. I'm not meaning to appeal to some sense of artistic integrity or "real photography" here. I just hate the lack of control over the product. I used to be really into photography as a hobby and had a lot of opinions about lenses and mostly set up exposures manually. Nowadays I use my Sony Alpha with the kit lens and rarely take it off of its "smart" auto mode, which does have some ML-driven features like subject detection. But it feels like I have so much more control over the output than I do with my phone, because the Sony doesn't run the image through ten layers of AI processing that's not a whole lot better than the state of the art in Instagram filters before saving it. If I don't hold the camera steady it'll just come out motion blurred, not like someone new to photoshop has just discovered the posterize button.
As I understand Apple is better than most of the Android vendors about this kind of thing and the iPhone processing probably produces better output - but it's still frustrating to me feeling like photos are changing from "capturing the scene" to "recreating the scene." I did graduate work on forensics of digital images, learned a lot of theory and methods for analyzing and reversing in-camera processing. I did some research on the "auto HDR" feature that was starting to appear in Android devices at the time and whether or not it defeated some known forensic methods for device fingerprinting (mostly, not totally). But that was the tip of the iceberg... it used to be that cameras only did a bit of processing, debayering for example, the kind of things that really need to be done to turn sensor data into a useful image because of the properties of the sensor and readout pipeline. But phones, the dominant photographic tool today, are taking it to this whole new level where they do what would have been very complex postprocessing on every image, as it's taken.
As with so many things, I guess it's good when it works, but endlessly frustrating when it doesn't. At least it feels like the phone vendors are doing their part to preserve "traditional" photographic technology, if that's what you'd call a Sony mirrorless, by really nerfing phones as tools for people who want much control over the result. I do understand there are third-party apps for iPhone that expose a lot more user control but it seems like they also have some limitations with how much of the camera stack they can control/bypass.
Yeah the whole Samsung moon shots (possibly r/Android's biggest moment as a subreddit) really kinda laid out, paired with the anxiety around AI, that our phone cameras...are not really capturing what we see anymore, or what was even there anymore. There's levels to it of course, but it is unsettling that we're going to be in this space of not even trusting any image for a long, long time.
I agree so much with this comment. Including feeling like I need to switch back to carrying around a proper camera.
When I switched to the Google Pixel 2, the post processing truely was revolutionary compared to other phone cameras and I stopped using anything other than the phone to take pictures. Even back then, iPhone had post processing turned up to a level that most pictures looked a bit "off" to me in the background details, but most people didn't seem to notice it, even if I tried to point it out. Google's flagship camera seemed to avoid that over processing and the results were really good. Unfortunately, Google seemed to get cocky about it and just kept increasing the level of processing as the years passed.
Now I'm on a Pixel 7 Pro that I got specifically for the 5x zoom camera and I've been consistently pissed off by every zoom picture. Even though it's an optical zoom, the processing gets turned up so high that I feel like it's worse than early days low resolution digital zoom. The picture basically looks a pretty decent prompt generated picture vs. a camera shot. It's kind of ridiculous how bad details get just made up with the pictures out of the 5x zoom lens. The 2x and 1x lenses are substantially better, but still frustrating.
The annoying thing is every photo looks pretty good in the preview thumbnail and even usually looks pretty good in the phone gallery. But if you zoom in or view on a monitor, the digital slurry in the background becomes to apparent. I haven't tried to print a photo from this phone, but I imagine they would generally look pretty poor.
They've optimized so incredibly heavily for the common use case (browsing pictures on your phone) that they've forgotten everything else.
Every electronic camera is going to do processing. Not sure one should reject based on standard processing and not being certain. I disagree with the judges that the process matters when your judging a result not the process. They basically rejected it for being too good.