This company has been caught multiple times going back and changing the location of recorded shots in their system after the fact. I don't know why they're still in business.
That's strange. I would assume this would be a problem unusually well-suited to machine learning techniques. Law enforcement misuse and racial bias I can see, but inaccuracy? It's a triangulation problem mostly.
The ACLU had a pretty good article on it a few years ago. It seems the inaccuracy comes from the number false positives and the resulting aggressive police response.
Ah, of course. More human error. I should have known.
The system probably works great in military situations, which I believe is what it was designed for. In a dense city where sound can echo multiple times off various buildings and other structures? It probably gets things wrong quite often. Add in trigger-happy cops that don't know how to interpret the data and you have a recipe for disaster.
not if you consider fireworks, car misfires, echos and weird geometries, and the fact that supersonic bullets have a sonic boom that travels with it....
that and the ai was probably only trained in black neighborhoods so it thinks loud bass or black accents are required to be a positive? i dunno
all of those different noises have distinct soundwave profiles, and different geometries can be accounted for either in software or with strategic placement of mics. I'm convinced this would be a good ML project, if we could find a way of enforcing without police bias, which, good luck.
i don't think so... each neighborhood is shaped differently and will have an effect on the sound profiles...
maybe if you set it up and calibrated it by shooting guns all around the city (:
they have this in DC… coincidentally the day with the most “gunshots” is also the 4th of July when hoards of people are openly lighting off fireworks of all kinds in the street.
I guess I should have said "in principle it should be possible to distinguish these sounds" ~~because yeah a couple people saying stuff to shit on these systems' technology.~~
edit: I expressed myself very poorly last night. I meant to say, "I guess I should have said 'in principle it should be possible to distinguish these sounds' because people are making valid observations in the comments about the notorious failures of this product"
I wouldn’t assume a company like ShotSpotter uses modern machine learning techniques. It’s got a pitiful accuracy rate and the company was founded 28 years ago. They seem more like a company that hires people with connections rather than a company that hires AI experts and buys Nvidia H100 GPUs by the gross.
How TF do you racially bias a gunshot sound? That doesn't make any fucking sense...
I was imagining selective enforcement/response, or restricting use to areas where one race predominates, or behaving differently in response to mic reports in white neighborhoods than black. Standard cop shit.
edit: yeah pretty much
Ah okay, so the article as written is just written stupidly. It doesn't seem like the system itself is racially biased, so much as the cops using it.
Yeah now that you mention it, it is ambiguous in that way isn't it.
We actually have a local paper that suffered from what I believe is happening here. Cops give news papers and stuff deeper access to stuff on the regular. Our local paper here wrote something SLIGHTLY negative about one of the cops, and immediately had all their 'press' credentials revoked. They stepped back in line shortly after and issued an apology to the police department...
I'd be willing to bet that's what's happening here. They're toeing the line of saying the racial bias is the cops by making it ambiguous...
acab
Well they are placed in mostly minority communities but above that, a ShotSpotter tech admitted in court that they often change the analyst interpretations of what is and isn't a gunshot at the request of their police department customers which, keep in mind, have successfully been used in court as evidence of a crime.
There is also overwhelming data showing that the majority of their alerts lead to no arrests. The Chicago IG believes this demonstrates false positives where as ShotSpotter (who changed their name after some criticism to SoundThinking) says that people can fire a gun and leave no evidence in spite of police investigating and asking people that would have witnessed it (and there being no victim).
Furthermore, the way it works is that AI 'assists' a human who then determines if it is a gun shot and then attempts to triangulate the position it came from. From the trial I mentioned earlier, ShotSpotter determined that there was no gun shot but then changed the analysis at LEO request but in actuality was proven to be a helicopter....
Furthermore, ShotSpotter keeps the details of its methodologies and models a secret and has refused an independent audit from IPVM.
So with all of that, one could easily argue that ShotSpotter/SoundThinking is as biased as a police officer and that the evidence is purely subjective and non-transparent.
By putting them only in predominantly minority neighborhoods. And I do mean only.
The bias comes in because they installed them in areas with more gunshots, which happen to be areas with more minorities.
By not spending excessively on detectors uniformly distributed across the metropolitan area, they are targeting minorities.
targeting minorities is their whole job and always has been
they just need more microphones & probably a network of cameras on every street corner.
News
Welcome to the News community!
Rules:
1. Be civil
Attack the argument, not the person. No racism/sexism/bigotry. Good faith argumentation only. This includes accusing another user of being a bot or paid actor. Trolling is uncivil and is grounds for removal and/or a community ban. Do not respond to rule-breaking content; report it and move on.
2. All posts should contain a source (url) that is as reliable and unbiased as possible and must only contain one link.
Obvious right or left wing sources will be removed at the mods discretion. We have an actively updated blocklist, which you can see here: https://lemmy.world/post/2246130 if you feel like any website is missing, contact the mods. Supporting links can be added in comments or posted seperately but not to the post body.
3. No bots, spam or self-promotion.
Only approved bots, which follow the guidelines for bots set by the instance, are allowed.
4. Post titles should be the same as the article used as source.
Posts which titles don’t match the source won’t be removed, but the autoMod will notify you, and if your title misrepresents the original article, the post will be deleted. If the site changed their headline, the bot might still contact you, just ignore it, we won’t delete your post.
5. Only recent news is allowed.
Posts must be news from the most recent 30 days.
6. All posts must be news articles.
No opinion pieces, Listicles, editorials or celebrity gossip is allowed. All posts will be judged on a case-by-case basis.
7. No duplicate posts.
If a source you used was already posted by someone else, the autoMod will leave a message. Please remove your post if the autoMod is correct. If the post that matches your post is very old, we refer you to rule 5.
8. Misinformation is prohibited.
Misinformation / propaganda is strictly prohibited. Any comment or post containing or linking to misinformation will be removed. If you feel that your post has been removed in error, credible sources must be provided.
9. No link shorteners.
The auto mod will contact you if a link shortener is detected, please delete your post if they are right.
10. Don't copy entire article in your post body
For copyright reasons, you are not allowed to copy an entire article into your post body. This is an instance wide rule, that is strictly enforced in this community.