| ▲ | jqpabc123 8 hours ago |
| Used incorrectly will lead to errors. Only one small little problem --- there is no way to tell if you are using it "correctly". The only way to be sure is to not use it. Using it basically boils down to, "Do you feel lucky?". The Fargo police didn't get lucky in this case. And now the liability kicks in. |
|
| ▲ | nkrisc 8 hours ago | parent | next [-] |
| Some basic investigatory police work (the kind they did before AI) would have revealed the mistake before an innocent woman’s life was destroyed. |
| |
| ▲ | jqpabc123 7 hours ago | parent | next [-] | | Yes. But doing the investigation negates much of the incentive for using AI. Look for similar to play out elsewhere --- using unreliable tools for decision making is not a good, responsible business plan. And lawyers are just waiting to press the point. | | |
| ▲ | nkrisc 7 hours ago | parent | next [-] | | In this case it sounds as though AI could have been used to generate preliminary leads. When someone calls a tip line with information, police don’t just take their word for it, they investigate it. They know that tips they receive may be incorrect. They should have done the exact same here, but they didn’t. I’m very opposed to AI in general, but this one is clearly human failure. The noteworthy AI angle is the undeserved credence police gave to AI information. But that is ultimately their failure; they should be investigating all information they receive. | | |
| ▲ | jqpabc123 7 hours ago | parent [-] | | ...but this one is clearly human failure. Absolutely. The failure starts with tool vendors who market these statistical/probabilistic pattern searchers as "intelligent". The Fargo police failed to fully evaluate these marketing claims before applying them to their work. So in the same way that the failure rolled down hill, liability needs to roll back up. |
| |
| ▲ | bornfreddy 7 hours ago | parent | prev [-] | | AI can provide leads. Someone still needs to verify them and decide. | | |
| ▲ | jqpabc123 7 hours ago | parent [-] | | Generating and verifying bad leads costs money. Not verifying bad leads can cost much more. At some point, you have to decide if wasting good money on bad intel makes sense. |
|
| |
| ▲ | SpicyLemonZest 3 hours ago | parent | prev [-] | | The article says that the Fargo police claimed to have done "additional investigative steps independent of AI". (Perhaps they're lying, or did a poor job because they thought the extra steps were a formality.) |
|
|
| ▲ | jfengel 8 hours ago | parent | prev | next [-] |
| Now the "qualified" immunity kicks in. |
| |
|
| ▲ | zephen 5 hours ago | parent | prev [-] |
| Look, I'm generally considered AI's most vociferous detractor. But... > there is no way to tell if you are using it "correctly". This simply isn't true, at least in cases like this. I know common sense isn't really all that common, but why would you give more credence to an untested tool than an untested crack-addled human informant? The entire point of the informant, or the AI in this instance, is to generate leads. Which subsequently need to be checked. |
| |
| ▲ | jqpabc123 3 hours ago | parent [-] | | There is no "correct" way to use AI in order to avoid bad results. The only prudent approach is to assume all results are bad until proven otherwise. But this approach negates much of the incentive to pay for questionable results. | | |
| ▲ | zephen 32 minutes ago | parent [-] | | > The only prudent approach is to assume all results are bad until proven otherwise. As is true with results from people. > But this approach negates much of the incentive to pay for questionable results. I'm not sure that follows. Even the crack-addled human informant has always been paid for questionable results. |
|
|