• Dojan@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Advocates: take survivors of abuse seriously.
    Society: Let’s have computers tell us what to do!

    I mean I guess the risk of repeated murder-suicide is pretty low…

  • Uriel238 [all pronouns]@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    In the late 1970s (I was a kid) the computer is always right was a common sarcastic parody of all the people who actually believed it.

    We’d discoverin the 1980s it was possible to have missing data, insufficient data or erroneous data.

    • turmacar@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      It’s a sentiment at least as old as the first things that we now call computers.

      On two occasions I have been asked, “Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?” … I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question.

      —Charles Babbage

    • masquenox@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Oh, it’s far worse than that… the value of our lives have been determined by the (so-called) “free market” for a very long time now.

      The machine is simply going to streamline the process.

  • Lvxferre@mander.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    The algorithm itself is just a big “whatever”. The key issue here is that some assumptive piece of shit decided to conclude, based on partial information, that those women would be safe in the future.

  • barsoap@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    3 months ago

    The way to use these kinds of systems is to have the judge came to an independent decision, then, after that’s keyed in, the AI spits out theirs and whichever predicts more danger is then acted on.

    Relatedly, the way you have an AI select people and companies to get spot-checked by tax investigators is not to show investigators the AI scores, but mix in AI suspicions among a stream of randomly selected people.

    Relatedly, the way you have AI involved in medical diagnoses is not to tell the human doctor results, but suggest additional tests to be made. The “have you ruled out lupus” approach.

    And from what I’ve heard the medical profession actually got that right from the very beginning. They know what priming and bias is. Law enforcement? I fear we’ll have to ELI5 them the basics for the next five hundred years.

    • Match!!@pawb.social
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      But that doesn’t save money and the only reason the capitalists want AI is saving money

    • madsen@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      I don’t think there’s any AI involved. The article mentions nothing of the sort, it’s at least 8 years old (according to the article) and the input is 35 yes/no questions, so it’s probably just some points assigned for the answers and maybe some simple arithmetic.

      • barsoap@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 months ago

        Sounds like an expert system then (just judging by the age) which was AI before the whole machine learning craze, in any case you need to take the same kind of care when integrating them into whatever real-world structures there are.

        Medicine used them with quite some success problem being they take a long time to develop because humans need to input expert knowledge, and then they get outdated quite quickly.

        Back to the system though: 35 questions is not enough for these kinds of questions. And that’s not an issue of number of questions, but things like body language and tone of voice not being included.

        so it’s probably just some points assigned for the answers and maybe some simple arithmetic.

        Why yes, that’s all that machine learning is, a bunch of statistics :)

        • madsen@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          3 months ago
          so it’s probably just some points assigned for the answers and maybe some simple arithmetic.
          

          Why yes, that’s all that machine learning is, a bunch of statistics :)

          I know, but that’s not what I meant. I mean literally something as simple and mundane as assigning points per answer and evaluating the final score:

          // Pseudo code
          risk = 0
          if (Q1 == true) {
              risk += 20
          }
          if (Q2 == true) {
              risk += 10
          }
          // etc...
          // Maybe throw in a bit of
          if (Q28 == true) {
              if (Q22 == true and Q23 == true) {
                  risk *= 1.5
              } else {
                  risk += 10
              }
          }
          
          // And finally, evaluate the risk:
          if (risk < 10) {
              return "negligible"
          } else if (risk >= 10 and risk < 40) {
              return "low risk"
          }
          // etc... You get the picture.
          

          And yes, I know I can just write if (Q1) {, but I wanted to make it a bit more accessible for non-programmers.

          The article gives absolutely no reason for us to assume it’s anything more than that, and I apparently missed the part of the article that mentioned that the system had been in use since 2007. I know we had machine learning too back then, but looking at the project description here: https://eucpn.org/sites/default/files/document/files/Buena practica VIOGEN_0.pdf it looks more like they looked at a bunch of cases (2159) and came up with the 35 questions and a scoring system not unlike what I just described above.

          Edit: I managed to find this, which has apparently been taken down since (but thanks to archive.org it’s still available): https://web.archive.org/web/20240227072357/https://eticasfoundation.org/gender/the-external-audit-of-the-viogen-system/

          VioGén’s algorithm uses classical statistical models to perform a risk evaluation based on the weighted sum of all the responses according to pre-set weights for each variable. It is designed as a recommendation system but, even though the police officers are able to increase the automatically assigned risk score, they maintain it in 95% of the cases.

          … which incidentally matches what the article says (that police maintain the VioGen risk score in 95% of the cases).

  • ✺roguetrick✺@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Despite this article, I’m still not convinced that the algorithms aren’t better. The policy states that people need to use their best judgement and can override the algorithms. The article argues that the algorithms are being over relied on. The article mentions in passing, however, that the statistics were worse before the algorithm was introduced.

    The point of the matter is, best judgement can be shitty. Your average cop has no idea what questions to ask without a list and how important they are per research. Some suggestions are too continue using the tool but use things like psychologists to administer it. The only way you could reasonably have a psych on call for every police station is to make it a remote interview, which frankly doesn’t seem better to me.

    In the end, the unstated problem is resources and how best to utilize them to prevent the violence. I’m sure Spain’s policy could be improved but shoring it up with an algorithm is a good practice.

    • ✺roguetrick✺@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Algorithms aren’t AI. They’re standardization measures in cases like this. Hell you don’t even need computers for many of them. We use tons in healthcare to classify risk, decide on treatment options, and even decide on how much medication to give.

  • Wispy2891@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    The computer response should be treated as just an indication and in all cases a human needs to decide to override that

    Otherwise we’ll all become useless pieces of a simulation

    I went to the bank to ask a loan and then it got rejected because the computer said I didn’t met the parameters by just 40 euro. Ah ok, I told the clerk, just lower the amount that I’m asking or spread it over a longer period. No, because after the quote is done and I signed the authorization for the algorithm to perform credit score, it can’t do it again in 3 months. What?? Call a supervisor and let them override it, 40 euro is so minimal that it’s not that big issue. No, impossible. So that means each single employee in the bank is just an interface to the computer and can be fired at will?

  • Leate_Wonceslace@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Pedantic Mathematician here.

    If it failed, then it was a heuristic, rather than an algorithm.

    Clearly, that’s the most important thing about this post.

    You’re welcome.

    • silence7@slrpnk.netOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Pretty much anything trying to predict human behavior is a heuristic; people using them as if they’ve got some kind of certainty is a problem.

    • technocrit@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      3 months ago

      Why not both? A bad algorithm based on bad heuristics? There are many many algorithms that fail at what they’re supposed to do.

      As a non-condescending “mathematician”, I’m happy to help.

  • IsThisAnAI@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    I have no issues with using ML to predict outcomes. It’s going to be wrong sometimes, so will humans. The system just needs review and input from humans understanding the model.

  • NutWrench@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    The police accepted the software’s judgment and Ms. Hemid went home with no further protection.

    This is what happens when you rely on your Nintendos, instead of using your damn brains.

    • aport@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Even when given the best and most sophisticated tools and equipment available, police will manage to fuck things up at every opportunity because they’re utterly incompetent.

    • rottingleaf@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      And that’s why I’m against ALL such things.

      Not because they can’t be done right and you can’t teach people to use them.

      But because there’s a slippery slope of human nature where people want to offload the burden of decision to a machine, an oracle, a die, a set of bird intestines. The genie is out and they will do that again and again, but in a professional organization, like police, one can make a decision of creating fewer opportunities for such catastrophes.

      The rule is that people shouldn’t use machines above their brains, as one other commenter says, and they should only use this in a logical OR with their own judgment made earlier, as another commenter says, but the problem is in human nature and I’d rather not introduce this particular point of failure to police, politics, anything juridical and military.

        • rottingleaf@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          3 months ago

          Cops are still necessary. It’s giving humans a machine to blame any failure upon is a very bad thing.

          I personally think these "AI"s are supported by governments. There’s been a lot of talk 10-15 years ago how many government official’s functions can be replaced by AI (without quotes), since these functions do not require agenda and are not even too fuzzy, but require semantic understanding. So "AI"s (with quotes) are being used like a vaccine, so that the wide mass of humans would hate the guts of the very idea, having experienced them (EDIT: and wouldn’t want actual semantic reasoning systems). Why - because people working in governments love power and hate transparency, they also hate the idea of being replaced with machines.

          Or maybe it’s a conspiracy theory and they all really believe in accelerationism.

          • Match!!@pawb.social
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 months ago

            some political groups engage in mismanagement on purpose to make people dislike the government, that’s hardly a conspiracy, but it’s a little weird to think they’re propping up the misuse of LLMs rather than that being a natural consequence of stupid capitalism

    • EatATaco@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      3 months ago

      But the system seems to be better than police officers. Which is entirely believable. Humans have all kinds of biases that make the decisions we make far less than desirable.

      Per the article, it has decreased the risk of repeated violence and, according to an expert, its the best systen we have. Why would you want to go back to a worse system? This is using our brains in an attempt to overcoming our biases.