The case adds to a growing number of wrongful facial recognition arrests.
Retailers increasingly are using facial recognition software to patrol their stores for shoplifters and other unwanted customers. But the technology’s accuracy is highly dependent on technical factors — the cameras’ video quality, a store’s lighting, the size of its face database — and a mismatch can lead to dangerous results.
Faulty software and people who don't operate it properly, this is going to cause a lot of problems in the coming years. I read another article about faulty Fujitsu auditing software in the UK that led to 400 postmasters being falsely accused of theft.
That shit was wild. It went on for so long and impacted so many people. At that point I have to think people were intentionally ignoring it. It ruined so many lives just to say "haha sorry, was a software glitch" when they could have investigated it at any point and definitely should have with the strange uptick in "embezzlement".
It was investigated. They opened an internal investigation and when it became obvious the investigator was going to blow it open, they cancelled the investigation 2 days before the report was going to come out.
Seriously, it's really looking like the entire board over several years should face jail time.
Is this not a case of an unavailable accusers? In other words, Traffic Cams have time and again been overturned as unconstitutional because you have the right to face your accusers?
Pretty sure Tom from IT can rustle up a laptop to take to court but...Facial Recog. is some bullshit 1984 bullshit.
Macy's, Sunglass Hut, Rite-Aid... We need a website with a list of the companies that falsely accuse their visitors of crimes. The public needs to know the dangers of shopping at these places.
Also, AI should be excluded as evidence of a crime. The results of any form of AI analysis should be forbidden in probable cause statements.
Agreed. If a store uses facial recognition, it can only be to falsely accuse their visitors of crimes. The only possible outcomes of facial recognition are unreliable results and false accusations.
Wild that you can base a whole case on what a photo AI thinks it is seeing. These programs at the very least should work like DNA or fingerprint matching and provide a percentage of its accuracy, not just that it finds some kinda close image in its database and everyone rolls with it. And it should need some other piece of evidence as well to back it up, it should never be the "best" part of a prosecutors case.
This is why traffic cams in the US have had issues for years, and most of them are run privately, and issue "civil fines".
Because "civil fines" (taxes, under another name, same as "civil fees") don't have the legal issues of receiving a ticket.
Tickets generally require interaction with an officer. Since cameras and their companies aren't officers, they can't generate a ticket/summons. So the gov end-runs this by using civil fees/fines, with the camera operators receiving upward of 85% of the fee.
And being a fee/fine, it's difficult to get out of, even if you're innocent and pursue it in court.
Of course, every jurisdiction is different, so it depends on the local legal structure.
To add onto this, here's a story about how someone who had their car stolen (and they could prove it) lost their initial objection to the charges from a red light camera.
With most digital forensic tools thats exactly what they do. There's a specific threshold that gives a match probability. It's designed as a way to point someone in a direction, not to confirm identity.
I can totally see cops using this as probable cause but it would get totally laughed out of a courtroom.
That dude deserves a payday at the very least. Facial recognition isn’t reliable yet. And even when it is that shit should still be illegal to use as the sole justification in arrests. Do some god damn police work. Especially if the dude has an air tight alibi like BEING IN JAIL. It would take 5 minutes to confirm that.
A man was sexually assaulted in jail after being falsely accused of armed robbery due to a faulty facial recognition match, his attorneys said, in a case that further highlights the dangers of the technology’s expanding use by law enforcement.
Harvey Murphy Jr., 61, said he was beaten and raped by three men in a Texas jail bathroom in 2022 after being booked on charges he’d held up employees at gunpoint inside a Sunglass Hut in a Houston shopping center, according to a lawsuit he filed last week.
A representative of a nearby Macy’s told Houston police during the investigation that the company’s system, which scanned surveillance-camera footage for faces in an internal shoplifter database, found evidence that Murphy had robbed both stores, leading to his arrest.
The company said in a previous statement that it uses “facial recognition in conjunction with other security methods in a small subset of Macy’s stores with high incidences of organized retail theft and repeat offenders.”
But the technology’s accuracy is highly dependent on technical factors — the cameras’ video quality, a store’s lighting, the size of its face database — and a mismatch can lead to dangerous results.
The Federal Trade Commission last month said the pharmacy chain Rite Aid had misused its facial recognition system in a way that led to shoppers being falsely accused of theft, including in confrontations with police.
The original article contains 805 words, the summary contains 230 words. Saved 71%. I'm a bot and I'm open source!