this post was submitted on 24 Jul 2023
100 points (100.0% liked)

Technology

37717 readers
404 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
top 45 comments
sorted by: hot top controversial new old
[–] I_like_cats@lemmy.one 69 points 1 year ago (2 children)

Great! Automated discrimination against the poor

[–] ram@lemmy.ramram.ink 21 points 1 year ago (2 children)

How long before we find out it always flags certain makes and models as criminals?

[–] SenorBolsa@beehaw.org 12 points 1 year ago* (last edited 1 year ago)

it will just hand out random DUIs to RAM owners based on statistics and reckless driving tickets to black altimas with tinted windows and bedazzled plate frames.

[–] Marsupial@quokk.au 6 points 1 year ago (1 children)

That would be great.

Imagine an AI that’s morally blind, pulling over expensive cars because the owners are more likely to break traffic laws knowing they can afford the fine.

[–] ram@lemmy.ramram.ink 3 points 1 year ago

Or an AI that's pulling over cheap and old cars because the owners are more likely to get ticketed due to living in over-policed neighbourhoods.

[–] ErwinLottemann@kbin.social 9 points 1 year ago* (last edited 1 year ago) (4 children)

do 'the poor' all drive really bad?

[–] lemillionsocks@beehaw.org 23 points 1 year ago (1 children)

We've already got numerous examples of how these ai models and face recognition models tend to have biases or are fed data that accidentally has a racial bias. Its not a stretch of the imagination to see how this can go wrong.

Yep, the age old "garbage in garbage out". If we had a perfect track record we could just send in all the cop data, but we know for a fact the poor and PoC are stopped more than others. You send that into AI it will learn those same biases

[–] snooggums@kbin.social 22 points 1 year ago (1 children)

No, but any automated system can be used to punish people who cannot afford to fight it.

[–] mrmanager@lemmy.today 3 points 1 year ago* (last edited 1 year ago) (1 children)

Well to be fair, this is because of the retarded justice system in the US.

Just the term "afford to fight for it" is something that never should exist in a civilized society.

[–] Lowbird@beehaw.org 2 points 1 year ago (1 children)

I agree, but can we not with the r word?

[–] mrmanager@lemmy.today 2 points 1 year ago

Yes I can edit it, didn't mean to offend.

[–] boonhet@lemm.ee 12 points 1 year ago

No, but they're disproportionately affected by fines. For rich people a fine is just the cost of a privilege.

Exception being something like Finnish speeding fines which are income dependent.

[–] MyTurtleSwimsUpsideDown@kbin.social 4 points 1 year ago* (last edited 1 year ago) (1 children)

Not that I’m aware of.
But we know the criminal justice system currently has biases. If the data the “AI” is trained on was affected by these biases, or others that we don’t realize, then it will produce biased results.

[–] Lowbird@beehaw.org 2 points 1 year ago

One of the worst parts about all this to me is that the AI and the dataset used to trained it are kept secret as proprietary information, and the police and governments buy it anyway despite that nobody can even try to check the code or dataset to see what biases or errors it might have (and definitely does).

[–] lemillionsocks@beehaw.org 24 points 1 year ago (2 children)

Wonder how long before these things start automatically mailing you a ticket because they calculated your speed between camera stops and determined you were going 3-5mph over the limit on the highway and if the "well it was used to catch a criminal" crowd will be as accepting of this technology then

[–] Xkdrxodrixkr@feddit.de 14 points 1 year ago (1 children)

The section control technology youre refering to is already being used in some countries, and they ofcourse automatically send you your speeding tickets. 3-5 mph would only be a marginal fine tho, if there even would be a fine at all. I dont understand though how that would be any different from normal speed checks, except for the fact that it might be more accurate.

[–] jmp242@sopuli.xyz 4 points 1 year ago

The problem in the US is they can't give moving violations without someone there to testify. Usually that's the officer. If the officer doesn't show up, the ticket is tossed. I'm not really sure why they can give redlight tickets (unclear if that held up or not), but some of it had to do with if it was something that affected your license, or was a "violation" instead of a crime like a parking ticket.

[–] teawrecks@sopuli.xyz 1 points 1 year ago

I mean, they can't prove a sufficiently large gravitational wave didn't hit you between the two checkpoints, causing you to travel through space relatively quickly.

[–] money_loo@1337lemmy.com 14 points 1 year ago (5 children)

Oh no, the A.I. identified someone as a drug trafficker, and the police pulled that person over on suspicion of being a drug trafficker, and found out that he was indeed a drug trafficker, and now he's upset he got caught by a robot dragnet.

I don't think drugs should be criminalized, but are we supposed to be upset that A.I. is going to finally help parse data and solve crimes?

[–] Dinodicchellathicc@lemmy.ml 42 points 1 year ago (2 children)

This is literally 1984. We shouldn't applaud when they propose a surveillance state

[–] money_loo@1337lemmy.com 5 points 1 year ago (1 children)
[–] lobelia581@lemmy.dbzer0.com 21 points 1 year ago (2 children)

There's a difference between using AI as a tool and using it as a solution. Though knowing how this society works, it'll start off as a tool like now, and soon enough the higher ups will wonder why humans are even necessary in the process, especially when they need to be paid, and against everyone else's objections they'll get rid of the human verification part and use only AI, and when things go wrong the people in charge will go "who could have seen it coming?"

[–] money_loo@1337lemmy.com 7 points 1 year ago

Sounds like oversight, transparency, and regulation are in order. There's no putting this genie back in the bottle, unfortunately.

[–] SSUPII@sopuli.xyz 2 points 1 year ago

Wasn't there recently the news of the suicide outline that fired its staff for generative AI chat bots until not even a few days later it started giving dangerous responses?

[–] money_loo@1337lemmy.com 2 points 1 year ago (1 children)

McDonalds and White Castle have already begun using ALPR to tailor drive-through experiences, detecting returning customers and using past orders to guide them through the ordering process

Yeah you're right, helping people order lunch is literally 1984.

Uh oh, the things you are buying look pretty suspicious, we are going to wiretap your house to make sure you are not doing any no-nos.

I really dislike people using technology to analyze my habits, I prefer just stumbling onto things I like because they were around the places I usually look. Yes, I also don't like algorithmic content, it just makes people try to appease the algorithm, meaning less effort into the thing they do

[–] webghost0101@sopuli.xyz 41 points 1 year ago* (last edited 1 year ago)

This time they where right because it was indeed a drugsdealer but just look at what it took to get this data

“in this case it was used to examine the driving patterns of anyone passing one of Westchester County’s 480 cameras over a two-year period.”

“the AI determined that Zayas’ car was on a journey typical of a drug trafficker. According to a Department of Justice prosecutor filing, it made nine trips from Massachusetts to different parts of New York between October 2020 and August 2021 following routes known to be used by narcotics pushers and for conspicuously short stays.”

So apparently making long trips with short stays is now enough proof to be searched by police. And if they can extrapolate that into “this guys a dealer” how much other data and possible extrapolations got caught in the crossfire off all those cameras. How long till someone in power decides selling some of that info to corporations is a good way to line state/government/own pockets?

Maybe we should place cameras in everyones house and listening bugs in every single phone? Criminality solved? Or hear me out, the real criminals will adept, find new and novel ways while the common citizen is kept in line with fines for even the smallest offense.

Then the police state will want to escalate the tools again, even more suppressing technology. Good thing were spend so much resources continuing to bully normal citizens into generating cash flow from fines. Money and resources well spend?

Or maybe the world needs some actually intelligent people that can find the root causes of criminal behavior and restructure society to improve well being and chances so people want to belong and maintain it rather then feeling like the system of opportunities is rigged against them so they should cheat to survive.

[–] Catoblepas@lemmy.blahaj.zone 33 points 1 year ago (1 children)

Acknowledging this is a dragnet, a practice generally considered unconstitutional since the 1950s, actually illustrates pretty well why people are upset about it. Even if it would result in more easy prosecutions for cops, it doesn’t change that it’s mass surveillance and an unconstitutional practice.

[–] money_loo@1337lemmy.com 2 points 1 year ago

I guess that will be up to the courts to decide, not us.

[–] MagicShel@programming.dev 29 points 1 year ago

This pattern might indicate drugs. Or adultery, which isn't illegal. It could be a straight job such as a mobile MRI technician. It might be a landlord.

In short this is likely to affect innocent people. It's like if you've got a name that happens to be on the no fly list, right? Your travel is fucked and you haven't done anything wrong.

[–] allywilson@sopuli.xyz 16 points 1 year ago (1 children)

I think people get worried about concepts like pre-crime with this.

[–] money_loo@1337lemmy.com 4 points 1 year ago (3 children)

Worry, what...that you'll get pulled over for suspicions while being innocent, and then the cops would be forced to find out you're innocent?

Yeah I can see that being inconvenient to downright dangerous depending on the cop, personally I think it has the potential to do more good than bad.

It's just a shame the police have such a hard on for drugs when there's so much worse stuff going on out there.

[–] Dr_Cog@mander.xyz 15 points 1 year ago

And a certain percentage of innocent people are found guilty. You don't see how expanding the arrests of innocent people is a bad thing? It has the potential to ruin lives

[–] programmer_belch@lemmy.dbzer0.com 15 points 1 year ago (1 children)

What about innocent until proven guilty?

Sir, we have proof you look like a drug dealer, you have to prove you are not one

[–] money_loo@1337lemmy.com 3 points 1 year ago (1 children)

Wait until you grow up a bit and learn about probable cause, or people getting pulled over for nothing. This tech changes nothing about your idea of “innocent until proven guilty” vs what actually happens on the daily.

[–] elfpie@beehaw.org 17 points 1 year ago

I think that's exactly the point. The current situation is already bad, tools that reinforce the bad part of the system shouldn't be accepted.

[–] N0_Varak@lemm.ee 7 points 1 year ago

Big "if you're innocent, you've got nothing to hide" energy. I can be innocent and still not want unnecessary interaction with police and appreciate my privacy

[–] RagingNerdoholic@lemmy.ca 10 points 1 year ago

I thought cops already had a system for that called "driving while black."

Oh neat, this knowledge will now passively motivate me so hard to act normal that it's gonna flag me every time!

[–] nieceandtows@programming.dev 5 points 1 year ago (2 children)

Wake me when the AI can identify people who are going to commit crimes before they commit the crimes.

[–] Magnus@lemmy.dbzer0.com 3 points 1 year ago

Why would you need to be woken up if you have no intent of committing a crime?

[–] rammer@sopuli.xyz 2 points 1 year ago

Better wake up then. Because they are using such systems in the US and they are used to harass innocent people.

[–] Magnus@lemmy.dbzer0.com 3 points 1 year ago (1 children)

Wait so if you drive like a crazy person after a few beers and a good ol'smash up with the lads then you are suddenly a criminal?

[–] nzodd@beehaw.org 4 points 1 year ago

I do my best driving after a few beers.

Granted, that's mostly because I have a much easier time hitting pedestrians when I'm sober.

load more comments
view more: next ›