Police used AI facial recognition to wrongly arrest TN woman for crimes in ND (cnn.com)

by ourmandave 204 comments 443 points
Read article View on HN

204 comments

[−] firefoxd 47d ago
Without even looking at the AI part, I have a single question: Did anybody investigate? That's it.

Whether it's AI that flagged her, or a witness who saw her, or her IP address appeared on the logs. Did anybody bothered to ask her "where were you the morning of july 10th between 3 and 4pm. But that's not what happened, they saw the data and said "we got her".

But this is the worst part of the story:

> And after her ordeal, she never plans to return to the state: “I’m just glad it’s over,” she told WDAY. “I’ll never go back to North Dakota.”

That's the lesson? Never go back to North Dakota. No, challenge the entire system. A few years back it was a kid accused of shoplifting [0]. Then a man dragged while his family was crying [1]. Unless we fight back, we are all guilty until cleared.

[0]: https://www.theregister.com/2021/05/29/apple_sis_lawsuit/

[1]: https://news.ycombinator.com/item?id=23628394

[−] rcvassallo83 47d ago
The thing about the legal system is there's no incentive to investigate to find the truth.

The incentive is to prosecte and prove the charges.

Speaking from the experience of being falsely accused after calling 911 to stop a drunk woman from driving.

The narrative they "investigated" was so obviously false, bodycam evidence directly contradicted multiple key facts. Officials are interested only seeking to prove the case. Thankfully the jury came to the right verdict.

[−] FireBeyond 47d ago
There’s a judge down in Texas, Dallas area I believe, who is in social media a lot because he will excoriate prosecutors who bring bs in to his court room. He’s not soft on crime but hard on rights and process. If a defendant did the wrong thing, he will have the appropriate amount of sympathy, down to zero. At times he will tell them, we all know you got lucky here, do better. But he won’t let prosecutors slate by on garbage charges or statements or investigations by police. Which leads to my primary point at least for this discussion in particular:

To me the scariest part of this as a process is how many times (I’d casually estimate at least 75%) it is blindingly obvious that the prosecutor has not read the statement of charges or officer statements until everyone is in front of the judge. I get on one hand this judge seems to often be handling probable cause hearings but so many of these should never have resulted in any paperwork being turned in to the prosecution, let alone anyone having to show up in court.

[−] rcvassallo83 45d ago
Scary process is an understatement, especially because I was facing a domestic violence charge.

Long story short, emotional abusive partner got drunk and verbally combative, despite my attempts to de-escalate. When nothing worked I went into my bedroom and locked the door. She started pounding on the door and demanded her things. I gave them to her and told her she needs a ride home she no longer welcome. She verbally abused and provoked me for 10 minutes before getting in her car. Took the keys, called 911. She grabbed me causing us both to fall a few minutes before the cops arrived and told them I threw her to the ground. We both had a couple scapes so they arrested us both.

Interfered with the 911 call, filed a false police report, assaulted me, caused property damage. She got charged with class c assault only and a dismissal. I felt like I was seen as guilty until proven innocent.

Fortunately I recorded all her verbal abuse (prosecution tried to use it against me and brought in DV expert to explain both her conduct and my 911 call as typical in IPV cases)

Fortunately the jury didn't buy it. I was literally being threatened with violence in my own home for telling her to leave. Between that and the bodycam statements contradicting her testimony I was shocked that they didn't drop the charges or offer a favorable plea deal.

The judge was absolutely fair, prosecutor bent on punishment, alleged victim was attempting to ruin my life (as captured in my audio)

Whew!

In the end the claims were so obviously fabricated that my attorney made no defense. It was clear that the accuser was not credible.

Perjury was provable. No consequences for her. This happened in Brazoria county

[−] phonon 47d ago
[−] komali2 47d ago
It's fascinating to me that judges are elected in Texas, and what's more, run as members of a political party.
[−] retrochameleon 47d ago
There needs to be consequences for shitty, procedure-ignoring police work. Period.

Minimum 1 year of jail time for grossly wrongful arrests that could be avoided with standard procedure or investigation tactics that were not applied.

[−] helterskelter 47d ago
I agree with this sentiment but when you start punishing this sort of thing you create more incentive to cover it up. It's a tricky problem and I'm not sure there's a perfect solution.

What we really need is a change in police culture.

[−] retrochameleon 47d ago
Then the system should be redesigned such that transparency is a priority and cover ups are not feasible. And when cover ups eventually get found out, the punishments even more severe.
[−] true_religion 47d ago
We already have administrative punishments for the police when they incorrectly assign blame and cause a public relations mess.

Is the termination of your career and/or potential retraining and social embarrassment not already an incentive to cover up?

[−] jcgrillo 47d ago

> change in police culture

until then, there's a simple rule which works well: never talk to a cop. Or at least say the minimum number of words possible, give them nothing to use against you. Present ID if they ask for it, but never admit anything. If they persist, "lawyer". That has worked for me.

[−] account42 46d ago
Add even more disincentives for coverups (i.e. hard prison time) and rewards for whistleblowers.
[−] vablings 46d ago
Medicine has a culture that adapts to this quite well. If you make an honest mistake and communicate it, you are often persecuted by your peers but not hung out to dry legally by your hospital and generally your actions are always defensible.

Similar practices are used in law enforcement, but the legal implications are seemingly more severe

[−] pstuart 47d ago
These dialogs always prompt me to chime in with my solution: make the police be self-insured, backed by their pension fund.

The police today have zero incentive to serve the public, they have zero skin in the game and can literally get away with murder.

Any time you hear the call for "law and order", that is the audience that supports the current system, because they like it like this.

[−] fc417fc802 47d ago

> The thing about the legal system is there's no incentive to investigate to find the truth.

The truth is much more complicated and involves politics. For example Seattle (and possibly other cities?) enacted a law that involves paying damages for being wrong in the event of bringing certain types of charges. But that has resulted in some widely publicized examples where the prosecutor erred by being overly cautious.

[−] bko 47d ago

> The narrative they "investigated" was so obviously false, bodycam evidence directly contradicted multiple key facts. Officials are interested only seeking to prove the case. Thankfully the jury came to the right verdict.

I don't get it, if they only care about prosecuting and proving the case, wouldn't they go by the bodycam evidence? They didn't prove the case. Maybe if their incentive was to prosecute and prove the charges, they'd go by the obvious evidence. Or am I missing something here?

[−] pylua 47d ago
There is an incentive . It’s called fraud by negligence. I’m hoping she sues everyone here.

That’s seems to be in the realm of poissibility here if I am understanding things correctly (imo)

[−] hnuser123456 47d ago
I would absolutely never call the police on a woman. Simply walk far away and let her be someone else's problem.
[−] belorn 47d ago
Society went through the necessary lessons with DNA and fingerprints. Putting people in jail because the computer produce a match is a terrible idea, especially when its done by an proprietary dark box that no one really understand why it claims there is a match. It can be used as a tool of investigations to give the investigators an hint to find real more substantial clues, but using it like in fiction where the computer can act as the single truth is terrible for society and justice.

A month ago or so people on HN discussed facial recognition when looking victims and perpetrators in child exploitation material, and people were complaining that meta did not allow this fast enough. Neither the article or the people in that discussion draw any connection that the issues in this article could happen. People seemingly want to think that the lesson is "Never go back to North Dakota", as that is a much easier lesson than considering false positives in detection algorithms and their impact on a legal system that is constrained in budget, time, training and incentives.

[−] latexr 47d ago
Yes, of course someone should have investigated, but the larger point here is that people don’t because they are being sold a false narrative that AI is infallible and can do anything.

We could sit here all day arguing “you should always validate the results”, but even on HN there are people loudly advocating that you don’t need to.

[−] bl4ckneon 47d ago
I think you missed many important points.

"The trauma, loss of liberty, and reputational damage cannot be easily fixed,” Lipps' lawyers told CNN in an email.

That sounds a LOT like a statement you make for before suing for damages, not to mention they literally say "Her lawyers are exploring civil rights claims but have yet to file a lawsuit, they said."

This lady probably just wants to go back to normal life and get some money for the hell they put her in. She has never been on a airplane before, I doubt she is going to take on the entire system like you suggest. Easier said than done to "challenge the entire system", what does that even mean exactly?

[−] themafia 47d ago

> Whether it's AI that flagged her

It absolutely was. There's no question of this. Now we need to ask how was the system marketed, what did the police pay for it, how were they trained to use it?

> anybody bothered to ask her "where were you the morning of july 10th between 3 and 4pm.

Legally that amounts "hearsay" and cannot have any value. Those statements probably won't even be admissible in court without other supporting facts entered in first.

> we are all guilty until cleared.

This is not at a phenomenon that started with AI. If you scratch the surface, even slightly, you'll find that this is a common strategy used against defendants who are perceived as not being financially or logistically capable of defending themselves.

We have a private prison industry. The line between these two outcomes is very short.

[−] NL807 47d ago

>No, challenge the entire system.

Agree in principle. But people like her does not have the resources, financially and emotionally to go through the legal system again. Unless there are charitable lawyers who are willing to do it on her behalf for free.

[−] not2b 46d ago
Clearview again. ICE is using it too, and their people think it is an oracle that is always correct, so that when someone shows a passport card or a RealID showing that they are someone else, a US citizen or permanent resident, they are usually accused of having a fake ID. It's a flawed tool and it misidentifies people sometimes.
[−] tmpz22 47d ago
IANAL but AFAIK custodial interrogation triggers Miranda, lawyers, and those awful awful civil liberties we’re trying to get rid of.

Better just to apply Musk or Altman software to the problem and avoid it entirely.

[−] garethsprice 47d ago
The vendor they used, Clearview AI, does not allow you to request data deletion unless you live in one of the half-dozen states that legally mandate it.

https://www.clearview.ai/privacy-and-requests

I have suddenly becomes very interested in New York's S1422 Biometric Privacy Act.

[−] advisedwang 47d ago
For me the worst thing in this case is that a JUDGE signed off on an arrest warrant with only a clearview match linking Ms Lipps to the crime.

A judge and the warrant process are supposed to be the safeguard against police doing shady stuff (like relying on an AI hit to decide who commit a crime). But if the judges can't be bothered...

[−] tlogan 47d ago
This is a weak or misleading story about AI.

First, the detective used the FaceSketchID system, which has been around since around 2014. It is not new or uniquely tied to modern AI.

Second, the system only suggests possible matches. It is still up to the detective to investigate further and decide whether to pursue charges. And then it is up to court to issue the warrant.

The real question is why she was held in jail for four months. That is the part that I do not understand. My understanding is that there is 30-day limit (the requesting state must pick up the defendant within 30 day). Regarding the individual involved, Angela Lipps, she has reportedly been arrested before, so it is possible she was on parole. So maybe they were holding her because of that?

Can someone clarify how that process works?

[−] mitchbob 47d ago
Earlier discussion (405 comments):

https://news.ycombinator.com/item?id=47356968

[−] repiret 47d ago
This isn't the first time this month I've read about someone suffering consequences of mistaken identity after their facial recognition said they look like someone who committed a crime. I'm sure this is starting to happen at an alarming rate.

The fundamental problem is that among the 350 million people living in the United States, there are a lot of pairs of people who look pretty darn similar. It used to be impractical to ask a question like "who in the US looks like the person in this security footage", and so as a matter of practicality, once you found someone who looks like the suspect, you probably also have other evidence, even if it's pretty weak, linking them to the crime.

But with AI, you can ask "who in the US looks like this person", and so we need to re-calibrate what it means if all you know is that someone looks like a suspect. I am of the opinion that "looks like someone," in the absence of any other evidence, is reasonable suspicion, but not probable cause, that you are the person you look like. Reasonable suspicion is enough for the police to stop you on the street and ask for your ID, but not enough to arrest you. There are other data points that alone might not even be reasonable suspicion, but could be combined with "looks like someone" to make probable cause, such as "was near the place at the time the crime happened".

AI isn't really the problem, even whether or not the AI's determination that two people look alike is valid or reviewed by a human isn't the problem. The problem is assuming that because two people look alike they must be the same person, even if you have no other evidence of them being the same person.

[−] oopsiremembered 47d ago
Money quote from someone quoted in the article:

"[I]t’s not just a technology problem, it’s a technology and people problem."

I can't. I just can't.

[−] lucasfin000 47d ago
The actual scariest part isn't that the AI got it wrong... it's that nobody felt the need to verify the AI. A tip from an anonymous caller can get investigated and found out if its true or not, and a match from a facial recognition system apparently does not. People haven't built better investigative tools they've just built better ways to skip around the investigation.
[−] mememememememo 47d ago
Wow thought the bar for probable cause for an arrest warrant would be much higher. Especially to drag soneone from another state.
[−] internetguy 47d ago
Insane. Not even an apology. And they ask why we should respect the police.
[−] shevy-java 47d ago
So cops used AI to attempt to investigate a crime. But, there was no crime - the arrest was wrong. Why can cops excuse themselves here for delegating their responsibilities (protecting society, allegedly that is) onto software? AI may also be written by some corporations to "tweak" this or that, see this foreign-looking guy being more likely to be AI-investigated. This is like the movie Minority Report - but stupid. IMO the courts should conclude that cops should not be allowed to use AI without having a prior, independently verified objective reasoning for any investigation. This mass sniffing that is currently going on is very clearly illegal. The current orange guy does not care about the law; see flock cameras aka spy cameras employed by the government on all car drivers at all times.
[−] selcuka 47d ago
[−] RagnarD 47d ago
She should sue the city controlling that police department, into oblivion. Or at least to the absolute max she can get.
[−] chaostheory 47d ago
I’m starting to believe that the internet will become a Dark Forest soon.
[−] pratyushsood 45d ago
a single data point (facial match) would never pass a security assessment. AI-assisted decisions need verification chains.
[−] indigodaddy 47d ago
A lot of dumb shit happens in this arena, where if you had just one smart cop, it could have been prevented. Here’s one from 2023:

https://youtu.be/lPUBXN2Fd_E

[−] giardini 47d ago
This has been posted at least twice before on HN.
[−] jqpabc123 47d ago
AI is a liability issue waiting to happen. And this is just another example.
[−] jeremie_strand 47d ago
[dead]
[−] ch3 45d ago
[dead]
[−] mistM 47d ago
[dead]
[−] ValveFan6969 47d ago
[dead]
[−] casey2 47d ago
[flagged]
[−] renewiltord 47d ago
[flagged]