481
submitted 4 months ago by silence7@slrpnk.net to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] Norgur@fedia.io 91 points 4 months ago

I really have a hard time deciding if that is the scandal the article makes it out to be (although there is some backpedaling going on). The crucial point is: 8% of the decisions turn out to be wrong or misjudged. The article seems to want us to think that the use of the algorithm is to blame. Yet, is it? Is there evidence that a human would have judged those cases differently? Is there evidence that the algorithm does a worse job than humans? If not, then the article devolves onto blatant fear mongering and the message turns from "algorithm is to blame for deaths" into "algorithm unable to predict the future in 100% of cases", which of course it can't...

[-] nalinna@lemmy.world 28 points 4 months ago* (last edited 4 months ago)

Could a human have judged it better? Maybe not. I think a better question to ask is, "Should anyone be sent back into a violent domestic situation with no additional protection, no matter the calculated risk?" And as someone who has been on the receiving end of that conversation and later narrowly escaped a total-family-annihilation situation, I would say no...no one should be told that, even though they were in a terrifying, life-threatening situation, they will not be provided protection, and no further steps will be taken to keep them from being injured again, or from being killed next time. But even without algorithms, that happens constantly...the only thing the algorithm accomplishes is that the investigator / social worker / etc doesn't have to have any kind of personal connection with the victim, so they don't have to feel some kind of way for giving an innocent person a death sentence because they were just doing what the computer told them to.

Final thought: When you pair this practice with the ongoing conversation around the legality of women seeking divorce without their husband's consent, you have a terrifying and consistently deadly situation.

[-] rottingleaf@lemmy.world 6 points 4 months ago

the only thing the algorithm accomplishes is that the investigator / social worker / etc doesn’t have to have any kind of personal connection with the victim

This even works for people pulling the trigger. Following orders, sed lex dura lex, et cetera ad infinitum.

[-] nalinna@lemmy.world 4 points 4 months ago

Yep! For all the psych nerds, it's pretty much a direct lift of the Milgram Shock Experiment

[-] TimeSquirrel@kbin.melroy.org 9 points 4 months ago

Thank you, this is why I came to the Fediverse from Reddit.

[-] IsThisAnAI@lemmy.world 2 points 4 months ago

IMO this place is far more an echo chamber than Reddit. Both places have their share of team based opinions but reddits diversity IMO is better at surfacing it.

[-] RobotToaster@mander.xyz 9 points 4 months ago

An algorithm is never to blame, some pencil necked desk jockey decided the criteria to get help that was used to create the algorithm, the blame is entirely on them.

That said, I doubt it would make any difference if a human was in the loop. An algorithm is still al algorithm, even if it's applied by a human. We usually just call that a "policy" though. People have been murdered by the paper sea for decades before we started calling it "algorithms".

[-] sunzu@kbin.run 9 points 4 months ago

Critical thinking spotted, proper authorities have been notified.

We will fix you!

[-] Fubarberry@sopuli.xyz 7 points 4 months ago

It reminds me of the debate around self driving cars. Tesla has a flawed implementation of self driving tech, that's trying to gather all the information it needs through camera inputs vs using multiple sensor types. This doesn't always work, and has led to some questionable crashes where it definitely looks like a human driver could have avoided the crash.

However, even with Tesla's flawed self driving, They're supposed to have far fewer wrecks than humans driving. According to Tesla's safety report, Tesla's in self driving mode average 5-6 million miles per accident vs 1-1.5 million miles for Tesla drivers not using self driving (US average is 500-750k miles per accident).

So a system like this doesn't have to be perfect to do a far better job than people can, but that doesn't mean it won't feel terrible for the unlucky people who things go poorly for.

[-] technocrit@lemmy.dbzer0.com 12 points 4 months ago

Wow Tesla said that Tesla was safe!?!? This changes everything.

[-] ShittyBeatlesFCPres@lemmy.world 10 points 4 months ago

That report fails to take into account that the Cybertruck is already a wreck when it rolls off the assembly line.

[-] OhNoMoreLemmy@lemmy.ml 8 points 4 months ago

Unfortunately, this is bad statistics.

The Teslas in self driving mode tend to be used on main roads, and most accidents per mile happen on the small side streets. People are also much safer where Teslas are driven than the these statistics suggest.

[-] Fubarberry@sopuli.xyz 2 points 4 months ago

There's not much concrete data I can find on accident rates on highways vs non-highways. You would expect small side streets accidents to have lower fatality rates though, with wrecks at highway speeds to have much higher fatality rates. From what I see, a government investigation into how safe autopilot is determined there were 13 deaths, which is very low number given the billions of miles driven with autopilot on (3 billion+ in 2020, probably 5-10billion now? Just guessing here since I can't find a newer number).

But yeah, there are so many factors with driving that it's hard get an exact idea. Rural roads have the highest fatality rates (making up to 90% of accident fatalities in some states), and it's not hard to image that Tesla's are less popular in rural communities (although they seem to be pretty popular where I live).

But also rural roads are a perfect use case for autopilot, generally easy driving conditions where most deaths happen due to speeding and the driver not paying attention. Increased adoption of self driving cars in rural communities would probably save a lot of lives.

[-] sugar_in_your_tea@sh.itjust.works 6 points 4 months ago

Here's another quote further down:

Since 2007, about 0.03 percent of Spain’s 814,000 reported victims of gender violence have been killed after being assessed by VioGén, the ministry said. During that time, repeat attacks have fallen to roughly 15 percent of all gender violence cases from 40 percent, according to government figures.

“If it weren’t for this, we would have more homicides and gender-based violence,” said Juan José López Ossorio, a psychologist who helped create VioGén and works for the Interior Ministry.

So no, not a scandal, it seems it is helping, but perhaps could be better. At least that's my read.

[-] EatATaco@lemm.ee 4 points 4 months ago

Is there evidence that a human would have judged those cases differently?

It implies that a human would have been worse. Or at least that an average human would be worse, the ones making the decision.

[-] yesman@lemmy.world 3 points 4 months ago

The article is not about how the AI is responsible for the death. It's likely that the woman would have died in the counterfactual.

The question is not "how effective is AI"? The question is should life or death decisions be made by an electrified Oracle at Delphi. You must answer this question before "is AI effective" becomes relevant.

If somebody was adjudicating traffic court with Tarot cards, would you ask: well how accurate are the cards compared to a judge?

[-] surewhynotlem@lemmy.world 5 points 4 months ago

Decisions should be made by whomever or whatever is most effective. That's not even a debate. If the tarot cards were right more often than the judge, fire the judge and get me a deck. Because the judge is clearly ineffective.

You can't privilege an approach just because it sounds more reasonable. It also has to BE more reasonable. It's crazy to say "I'm happy being wrong because I'm more comfortable with the process"

The trick of course is to find fair ways to measure effectiveness accurately and make sure it's repeatable. That's a rabbit hole of challenges.

[-] rottingleaf@lemmy.world 1 points 4 months ago

The judge can bear legal responsibility. It's a feedback loop - somebody should be responsible for failures. We live in a society. If that somebody is not the side causing failures, things will get bad.

With a deck of cards it should be decided, how the responsibility is distributed between the party replacing humans with it, company producing cards, those interpreting the results.

[-] madsen@lemmy.world 4 points 4 months ago

Your point is valid regardless but the article mentions nothing about AI. ("Algorithm" doesn't mean "AI".)

[-] silence7@slrpnk.net 3 points 4 months ago

My impression from the article is more that they're not doing any kind of garbage-in assessment: nobody is making sure they're getting answers about the right person (eg: some women date more than one guy) and some women don't feel safe giving accurate answers to the police, and there aren't good failsafes available for when it's wrong; you're forced to hire legal counsel and pursue a change via the courts.

[-] nalinna@lemmy.world 6 points 4 months ago

That and, their action for low-risk is all wrong. The stakes are too high to not give someone help, regardless of the risk level.

[-] madsen@lemmy.world 1 points 4 months ago

The crucial point is: 8% of the decisions turn out to be wrong or misjudged.

The article says:

Yet roughly 8 percent of women who the algorithm found to be at negligible risk and 14 percent at low risk have reported being harmed again, according to Spain’s Interior Ministry, which oversees the system.

Granted, neither "negligible" or "low risk" means "no risk", but I think 8% and 14% are far too high numbers for those categories.

Furthermore, there's this crucial bit:

At least 247 women have also been killed by their current or former partner since 2007 after being assessed by VioGén, according to government figures. While that is a tiny fraction of gender violence cases, it points to the algorithm’s flaws. The New York Times found that in a judicial review of 98 of those homicides, 55 of the slain women were scored by VioGén as negligible or low risk for repeat abuse.

So in the 98 murders they reviewed, the algorithm put more than 50% of them at negligible or low risk for repeat abuse. That's a fucking coin flip!

[-] roguetrick@lemmy.world 1 points 4 months ago

You'll get that result without an algorithm as well unfortunately. A domestic violence interview often doesn't result in you getting the truth of what happens because the victim is often economically and emotionally dependent on their partner. It's helpful to have an algorithm that makes you ask the right questions but there's still no way I know of to get the right answers of those questions from a victim 100 percent of the time.

[-] madsen@lemmy.world 1 points 3 months ago

Odd. I replied to this comment, but now my reply is gone. Gonna try again and type up as much as I can remember.

Regardless, an algorithm expecting binary answers will obviously not take para- and extralinguistic cues into account. That extra 50 ms hesitation, the downwards glance and the voice cracking when answering "no" to "has he ever tried to strangle you before?" has a reasonable chance to get picked up by a human, but when reducing it to something that the algorithm can handle, it's just a simple "no". Humans are really good at picking up on such cues, even if they aren't consciously aware that they're doing it, but if said humans are preoccupied with staring into a computer screen in order to input the answers to the questionnaire, then there's a much higher chance that they'll miss them too. I honestly only see negatives here.

It’s helpful to have an algorithm that makes you ask the right questions [...]

Arguably a piece of paper could solve that problem.

Seriously. 55 victims out of the 98 homicide cases sampled were deemed at negligible or low risk. If a non-algorithm-assisted department presented those numbered I'd expect them to be looking for new jobs real fast.

[-] roguetrick@lemmy.world 1 points 3 months ago* (last edited 3 months ago)

I think beyond that it's purely the failure of the interviewer and not the tool. I think getting rid of the tool will just leave you with shitty interviewers and back to the same situation as you had before.

I've given plenty of algorithmic driven assessments myself, though mine are generally much shorter and the weights on the questions much simpler (plus I know the actual reasons behind the weight of my questions and why I'm asking them). You can always intervene when someone's lying and redirect them and can override the algorithm just like this Spanish policy. Lazy judges and police will exist without the tool.

It might be helpful for the tool to include a label that the interviewer thinks the result is unreliable due to the evasiveness of the interviewee, if only to show where the problems are coming from.

this post was submitted on 18 Jul 2024
481 points (96.5% liked)

Technology

59414 readers
1401 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS