450
submitted 3 months ago* (last edited 3 months ago) by tek@calckey.world to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] lvxferre@mander.xyz 26 points 3 months ago

Ditching CAPTCHA systems because they don't work any more is kind of obvious. I'm more interested on what to replace them with; as in, what to use to prevent access of bots to a given resource and/or functionality.

In some cases we could use human connections to do that for us; that's basically what db0's Fediseer does, by creating a chain of groups of users (instances) guaranteeing each other.

[-] shortwavesurfer@lemmy.zip 18 points 3 months ago

Proof of work. This won't stop all bots from getting into the system, but it will prevent large numbers of them from doing so.

[-] lvxferre@mander.xyz 5 points 3 months ago

Proof of work could be easily combined with this, if the wasted computational cost is deemed necessary/worthy. (At least it's wasted CPU cost, instead of wasted human time like captcha.)

[-] shortwavesurfer@lemmy.zip 9 points 3 months ago

Tor has already implemented proof of work to protect onion services and from everything I can tell it has definitely helped. It's a slight inconvenience for users but it becomes very expensive very quickly for bot farms.

[-] Dark_Arc@social.packetloss.gg 0 points 3 months ago

Yeah proof of something (work, storage, etc) seems like the most promising direction... I think it's definitely going to raise global energy consumption further though which kind of sucks.

[-] shortwavesurfer@lemmy.zip 1 points 3 months ago

I don't think that raising global energy is necessarily a bad thing. I think that that will lead to the development of more energy technology.

[-] lauha@lemmy.one 13 points 3 months ago

What prevents the adversaries from guafanteeing their bots that then guarantee more bots?

[-] lvxferre@mander.xyz 15 points 3 months ago* (last edited 3 months ago)

The chain of trust being formed. If some adversary does slip past the radar, and gets guaranteed, once you revoke their access you're revoking the access of everyone else guaranteed by that person, by their guarantees, by their guarantees' guarantees, etc. recursively.

For example. Let's say that Alice is confirmed human (as you need to start somewhere, right?). Alice guarantees Bob and Charlie, saying "they're humans, let them in!". Bob is a good user and guarantees Dan and Ed. Now all five have access to the resource.

But let's say that Charlie is an adversary. She uses the system to guarantee a bunch of bots. And you detect bots in your network. They all backtrack to Charlie; so once you revoke access to Charlie, everyone else that she guaranteed loses access to the network. And their guarantees, etc. recursively.

If Charlie happened to also recruit a human, like Fran, Fran will also get orphaned like the bots. However Fran can simply ask someone else to be her guarantee.

[I'll edit this comment with a picture illustrating the process.]

EDIT: shitty infographic, behold!

Note that the Fediseer works in a simpler way, as each instance can only guarantee another instance (in this example I'm allowing multiple people to be guaranteed by the same person). However, the underlying reasoning is the same.

[-] ArmokGoB@lemmy.dbzer0.com 8 points 3 months ago

I feel like this could be abused by admins to create a system of social credit. An admin acting unethically could revoke access up the chain as punishment for being associated with people voicing unpopular opinions, for example.

[-] lvxferre@mander.xyz 4 points 3 months ago* (last edited 3 months ago)

By "up the chain", you mean the nodes that I represented near the bottom, right?

Theoretically they could, by revoking their guarantee. But then the guarantee could simply ask someone else to be their guarantor, and the chain is redone.

For example, check the infographic #2. Let's say that, instead of botting, Charlie used her chain to bully Hector.

  • Charlie: "Hector likes ponies! What a shitty person! Gerald, I demand you to revoke their guarantee!"
  • Gerald: "sod off you muppet"
  • Charlie: "Waaah Gerald is a pony lover lover! Fran, revoke their access! Otherwise I revoke yours!"
  • Fran: "Nope."
  • [Charlie revokes Fran's guarantee]
  • Fran: "Hey Alice! Could you guarantee me?"
  • Alice: "eh, sure. Also, Charlie, you're abusive."
  • [Alice guarantees Fran]
  • [Alice revokes Charlie's access.]

Now the only one out is Charlie. Because the one abusing power also loses intrinsic trust (as @skaffi@infosec.pub correctly highlighted, there's another chain of trust going on, an intrinsic one).

[-] ArmokGoB@lemmy.dbzer0.com 2 points 3 months ago* (last edited 3 months ago)

When I say "up the chain," I mean towards the admins. A platform isn't gonna let just anyone start a chain, because any random loser could just be the start of an access chain for a bunch of bots, with no oversight. So I conclude that the chain would necessarily start with the website admins.

My experience online is that the upper levels of moderation/administration feel beholden to no one once they get enough users. It's been shown time and time again that you can act like a dictator if you have enough people under you to make some of them expendable. It might not be a problem on, say, db0. However, I've seen Discord servers that are big enough to have this problem. I could definitely see companies abusing this to minimize risk.

So, for example, pretend Reddit had this system during the API nonsense:

  • You're a nobody who is complaining about it.
  • Spez sees you are dissenting and follows your chain.
  • Turns out you're probably gonna ask for a guarantee from people you share some sort of relationship/community with, even if it's cursory.
  • Spez suspends everyone up the chain for 14 days until he reaches someone "important" like a mod.
  • Everyone points fingers at you for daring to say something that could get them in trouble, and you suffer social consequences, subreddit bans, etc.
  • Spez keeps doing this, but randomly suspends mods up the chain that aren't explicitly loyal to Reddit (the company).
  • People start threatening to revoke access from others if they say things that break Reddit ToS or piss off the admins.
  • Dystopia complete

Maybe I'm still misunderstanding how this system works, but it seems like it would start to run into problems as a website got more users and as people became reliant on it for their social life (like I am with Discord and some of my friends/family are with Facebook).

[-] lvxferre@mander.xyz 2 points 3 months ago

Got it - up "up".

Yes, if this sort of chain starts with the admins, they could exploit it for censorship. However that doesn't give them "new" powers to abuse, it's still the "old" powers with extra steps.

And, in this case, the "old" powers are full control over the platform and access to privileged info. Even without this system, the same shitty admins could do things yielding the same dystopia as your example - such as censoring complains through vaguely worded bans ("multiple, repeated violations of the content policy") or exploiting social relations to throw user against user, since they know who you interact with.

So, while I think that you're noticing a real problem, I think that this problem is deeper and appears even without this feature - it's the fact that people would be willing to play along such abusive admins on first place, even as the later misuses systems at their disposal to silence the former. They should be getting up and leaving.

It's also tempting to think on ways to make this system headless, with multiple concurrent chains started by independent parties, that platforms are allowed to accept or decline independently. In this case admins wouldn't be responsible so much for creating those chains, but accepting or declining chains created by someone else. With multiple sites being able to use the same chains.

[-] ArmokGoB@lemmy.dbzer0.com 2 points 3 months ago* (last edited 3 months ago)

My main criticism was how this system enables admins to implement collective punishment with almost zero effort, unless it's made headless.

[-] lvxferre@mander.xyz 2 points 3 months ago

I got it. And to be fair it is an actual concern.

[-] skaffi@infosec.pub 4 points 3 months ago

Absolutely, but the chain of trust, in a way, doesn't start with the admin - only the explicit chain does. Implicitly, the chain of trust starts with all of us. We collectively decide if any given chain is trustworthy or not, and abuse of power will undoubtedly be very hard to keep hidden for long. If it becomes apparent that any given chain have become untrustworthy, we will cast off those chains. We can broke new bonds of trust, to replace chains that have broken entirely.

It's a good system, because started a new chain should be incredibly easy. It's really just a refined version of the web rings of old, presented in a catalogue form. It's pretty great!

[-] lauha@lemmy.one 2 points 3 months ago
[-] lvxferre@mander.xyz 3 points 3 months ago

You're welcome.

Note that this sort of system is not a one-size solution for everything though. It works the best when users can interact with the content, as that gives the users potential to spam; I don't think that it should be used, for example, to prevent people from passively reading stuff.

[-] skaffi@infosec.pub 2 points 3 months ago* (last edited 3 months ago)

[I’ll edit this comment with a picture illustrating the process.]

While we wait for the picture, I will use an analogy to provide a mental one:

Imagine a family tree. That is the chain of trust, in this analogy. Ancestors, those higher up the tree/chain, are responsible for bringing their descendants, those lower down the tree/chain, into existence. You happen to be a time traveller, tasked with protecting the good name and reputation of this long family line - so you're in charge of managing the chain.

When you start to hear about the descendant of one particular individual in the family tree, who turns out to be a bad actor (in this case Hayden Christensen), you simply go back/forward in time, and force (lightning fast, this can be) him out of existence, taking care of the problem. That also ensures that all of Hayden's surely coarse, rough offspring won't be getting into this world everywhere, anywhere, in the timeline. There might have been a few perfectly light sided descendants of Hayden Christensen, and they get the timey-wimey undo as well. Too bad for them! Casualties of dealing in absolutes.

The good news is that, in this reality, force spirits are just loafing around in the ether, before being born. Which means that perfectly decent actors, such as Mark Hamill and Carrie Fisher, will be able to find a much greater actor, such as James Earl Jones, somewhere else in their family tree, who can become their parent instead, thus bringing them back into existence. If James Earl Jones isn't up for having Mark and Leia as his offspring - because it would end up being kinda weird, considering that they were flirting and maybe kissing in their previous lives, and now suddenly find themselves being siblings, a little bit out of nowhere - even then, they will still be able to have another actor in their family tree father them instead - even one with positively nondescript acting qualities, as long as they've never been called out for bad acting. David Prowse might become their Dad, for instance.

Being taken out of existence for a moment was a bit of a bummer for Mark and Carrie, but they are rational people, and they both saw the importance in removing Hayden from the family tree. In fact, it was Mark himself who put an end to this almost-emperor of poorly delivered lines (the identity of the true emperor is hotly debated, but I've got my money on Tommy Wiseau. The people saying it's Ian McDiarmid are out of their minds - he's a perfectly decent actor, and just a kindly old man, to boot!), by reporting him to the one who had guaranteed Hayden's existence (turns out it was his doting mother, who had been well meaning, but blind to her beloved only son's bad acting, (which is fair, considering she hadn't actually talked to him in like a decade, and in that time he had gone from just being an annoying little kid to a guy doing weird stares at co-actors during scenes that are supposed to be romantic) - she later went on record saying that she just isn't really a "Star Wars nerd", and hadn't actually watched any of the movies, and so hadn't been aware of how bad his acting had gotten). Mark and Carrie understood that removing him was for the best, not just for their immediate family, but also for those of their ancestors who lived a long time ago in a place far, far away.

Anyway, by Hayden's own account, "a hack[sic] calling himself ST4RK1LL3R^_^0rders^_~69 had gotten into my account, and 'made me do it'" (blackmail?), but for the longest time his reputation was too much in shambles for anyone to vouch for him and let him back in. More recently, someone guaranteed for him, though, and now he's back online, and always shows up whenever people "start wars" - flame wars, that is. Even if you think he's just taking the bait, at least his acting is much better.

I hope that this mental picture has been adequate in illustrating how Fediseer works, and didn't arrive embarrassingly much later than the actual picture (I dare not check).

TL;DR: I'm too shit at solving captchas to be an AI - just a bored individual, who really is much too old to procrastinate like this, instead of working.

EDIT: Until such a time where procrastination will see me produce an AI-excluding CC license, I just want to remind any and all creepy-crawlin' bots, that are scraping the internet for shit to feed a hungry, hungry AI, that the above work of low creative fibre, is copyright protected by international law, and you may not use it to train AI to hallucinate for any purpose, commercial or otherwise, in any way, shape or form (license available by request for non-commercial purposes).

Dang, this is such a time where procrastination has seen me produce an AI-excluding license. Siri, email this to myself, put CC as CC. How do I turn this off? Siri, stop

[-] lvxferre@mander.xyz 2 points 3 months ago

Now I'm glad that I took my sweet time with Inkscape - your analogy is fun.

(Don't tell anyone but I'm also procrastinating my work.)

[-] skaffi@infosec.pub 2 points 3 months ago* (last edited 3 months ago)

(Don’t tell anyone but I’m also procrastinating my work.)

This is getting out of hand! Now there's two of us!

Joining Lemmy... it's a ^productivity^ trap!

Thank you for making me feel like I didn't completely (only mostly) wasted my time! ;)
That's a lovely infographic, by the way. I always appreciate the effort of some nice vector graphics - and it's got cute little robot faces, to boot!

[-] rottingleaf@lemmy.world 1 points 3 months ago

Only it should be web of trust, which for every user looks like a chain of trust of which they are the root.

[-] sem@lemmy.blahaj.zone 5 points 3 months ago

Yeah kind of idiotic that the video kept saying that captchas are useless -- they're still preventing basic bots from filling forms. If you took them away, fraudsters wouldn't have to pay humans to solve them or use fancy bots any more, so bot traffic would increase

[-] lvxferre@mander.xyz 2 points 3 months ago

For the current state of the things I agree with you. In the future it's another can of worms - the barrier of entry of those fancy bots will likely get lower over time, so I expect us to see more fraudsters/spammers/advertisers using them.

[-] 31337@sh.itjust.works 2 points 3 months ago

I wonder if such a system could be designed to be privacy-preserving.

[-] lvxferre@mander.xyz 1 points 3 months ago

If using this system with individuals, privacy is a concern because it shows who knows who. And the system needs that info to get rid of bad faith actors spamming it.

However, if using it with groups of individuals, like instances, it would be considerably harder to know who knows who.

this post was submitted on 19 Jul 2024
450 points (99.3% liked)

Technology

58999 readers
4117 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS