14 out of 15 requests were of black people. Facial recognition is notoriously bad with darker skin tones.
Racial Discrimination in Face Recognition Technology https://sitn.hms.harvard.edu/flash/2020/racial-discrimination-in-face-recognition-technology/
Yeah, this same exact story keeps coming up for years now just with different names. Why anyone would think that both the ineffectiveness and racial bias in these systems either wouldn’t exist or will somehow go away eventually is beyond me. Just expensive and ineffective mass surveillance for the sake of it…
Actually, all 15 were of black people. 14 were of black men, one was a black woman.
Zero arrests as well.
New Orleans is pretty black, but thats just impressive.
Minor correction.
15 out of 15 requests were of black people. 14 of those requests were black men and 1 was a black woman.Thank you for your service!
deleted by creator
Yeah, but statistics is a b*tch.
We had a similar technology for a test run some years ago at a train station in Berlin, capital of Germany and largest city in the EU with 3.8M.
The results the government happily touted as a success were devastating. They had a true positive rate of 80% (and this was already cooked since they tested several systems but only reported the best results), which is really not that good to start with.
But they were also extremely proud of the false negative rate, which was below 0.1%. That doesn’t sound too bad, does it?
Well, let’s see…
True positive means you actually identified the people you were looking for. Now, I don’t know the number of people Berlin’s police is actively looking for, but it’s not that much. And the chances of one of them actually passing that very station are even worse. And out of that, you have 20% undetected. That’s one out of five. Great. If I were a terrorist, I would happily take that chance.
So now let’s have a look at the false negative rate, which means you incorrectly identified a totally harmless person as a terrorist/infected/whatever. The population for that condition is: everyone passing through that station.
Let’s assume there’s a 100k people on any given day (which IIRC is roughly half of what that station in Berlin actually has). 0.1% of 100k is 100 people, every day, who are mistakenly reported as „terrorists“. Yay.
How about 15/15?
Who remembers the HP computer that was unable to identify black people? One of my favorite “oooph, that’s not a good look” tech fails of all time. At least the people in that video were having a good laugh about it.
https://www.youtube.com/watch?v=t4DT3tQqgRM
Holy hell, that was 13 years ago.
Yeah. Basicly anything with a lower contrast, with shadows and backgrounds. And because shadows are dark, they have a lower contrast with other dark things.
Discrimination is the wrong word. Technology has no morals or sense of justice. It is bias in the data that developers should have accounted for.
It’s totally accurate though. It’s like the definition of systemic racism really. Think about housing or financial policy that disproportionately fails for minorities. They aren’t some Klan manifesto. Instead they just include banal qualifications and exemptions that end up at the same result.
This seems shortsighted. You are basically asking people to police their own biases. That’s a tall ask for something no one can claim immunity from.
Ask the people who create the data sets that machine learning models train on how they feel about racism and get back to us
It can be an imported bias/descrimination. I still think that words fair.
Do you have a more accurate word?
You need to learn some critical race theory. Racist systems turn innocent intentions into racist actions. If a PhD student trains an AI model on only white people because the university only has white students, then that AI model is going to fail black people because black people were already failed by university admissions. Innocent intention plus racist system equals racist action.
Even CRT would call this “racial bias”, which is exactly what this is.
Huh. It’s almost like cops are constantly wasting money on bullshit.
only if it’s ours, of course
The terrifying part to me is that cops across the nation have a long history of seeing that the tech they want to use is unreliable and based on junky science, but they still push it through anyway. Aren’t police dogs about as reliable as a coin-flip when their handlers aren’t nipping at their neck to get the to jump at anything? They don’t care if it’s right as long as they can use it to justify their behavior, so they make it policy.
Only the drug dogs are ineffective. Bloodhounds and tracking dogs have been a staple of hunting down people, and German retrievers can take a man down effectively as well.
When they are trained with incentives for finding something, instead of incentives to be correct, then they will find something. Same is true for man or beast.
A lot if forensic “science” is utter bunk. Yet it continues to be used. Having a fair and equitable system was never the point.
People may see this as a “see, AI isn’t that good”. We all need to rail against these kinds of programs to the point they are made illegal. Because there are examples around the world of being able to track people with facial recognition (and even by the way someone walks with their face entirely covered 0_0)
I see this as the new Orleans police dep hired a inept contractor (or did an inept job in house).
Around the world, we must fight against all inappropriate data harvesting.
With all the laws trying to put women into basically servitude I’m definitely on team rail against. There are a lot of types of “criminals” that need to be able to get away from law enforcement these days unfortunately. Honestly I’d prefer they just keep being inept for now lol
Tbf, NOPD don’t arrest many people anyway. There’s a massive cop shortage, only 944 officers for a city of 364,000 with skyrocketing crime rates. Moreover, they’ve been operating under a consent decree by the DOJ since 2012. They’re overworked, underpaid and under the thumb of the feds so in response they simply don’t do shit.
Serious question, what is the right number of officers for a city that size? 1 officer per 400 people or so doesn’t sound very low to me.
NOPD’s stated goal is 1600, a ratio of 1:227 persons.
The actual ratio is 1:385
Cleveland, similar in size to New Orleans, has a ratio 1:310. They also state that they are suffering from a serious police shortage.
By comparison:
NYC has a ratio of 1:166
Chicago 1:180
0 officers
Should be closer to 1 to 100. A large part of is overhead, desk clerks, admin, logistics guys, people like that. They all count as cops. And remember, they’re only working ine shift a day, most of the time they’re not working. And they usually end up in court for every ticket and arrest.
So yeah, they’re understaffed.
1:100 seems insane. I live in a community of 10,000 people, and we don’t have anything close to 100 people working in the RCMP. MAYBE 20-30, including support
Just checked our major urban centre in Canada, and it’s around 1:450. As a comparison, that makes New Orleans (1:385) pretty well staffed.
Would be cool to find data covering major urban centres across the world for comparisons.
New York City has a ratio of 1:166
And NYC is wildly over-policed.
I swear, some people have never met a societal problem they didn’t want to throw a cop at. Meanwhile we have more cops and prisoners per Capita than most of the world, funny how that works…
I know more than 100 people and zero criminals. Somehow this math doesn’t math to me.
The cops in my city were under a DOJ consent decree for like 20 years, and it didn’t make them any less effective. They’re actually worse now, because they actively don’t give a fuck.
Good.
deleted by creator
lots of nice biometric additions to the database tho, right? 😠
Only 15 additions actually
So rolling it out state- or nationwide next?
well, good on them for not arresting false positives at least
I mean, law enforcement occasionally uses polygraph tests in their investigations even though that type of “evidence” isn’t admissible in court and, to be honest, what kind of scientific credibility does a piece of technology like a polygraph even have? They’ll use whatever they can get their hands on even if it’s questionable. Some police forces probably even have a psychic consultant or something. It scares me.
They’ll use it especially if it’s questionable, like handwriting analysis, because the goal is arrests not correct arrests. Trumped up, flimsy, circumstantial “evidence” is the best kind when you don’t actually want to do your job.
When I walk into the building I work at there is a disclaimer that they are using facial recognition. I don’t know if this is reality or a scare tactic, but based on the industry I would assume they’re just using it for free AI training
The current state of policing doesn’t deserve to have access to this kinda shit. Hopefully it never will tbh.
if you think it’s good that cops have more tech you are the dumbest fucking hog imaginable
NOPD failing its citizen, one bad idea at a time