Deleted

  • jerkface@lemmy.ca
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    It’s not so important to tell the difference between a human and a bot as it is to tell the difference between a human and ten thousand bots. So add a very small cost to passing the test that is trivial to a human but would make mass abuse impractical. Like a million dollars. And then when a bot or two does get through anyway, who cares, you got a million dollars.

    • darkrai9292@lemmy.ca
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Yeah this seems to be the idea behind mCaptcha and other proof of work based solutions. I noticed the developers were working on adding that to Lemmy

  • CanadaPlus@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    Any bot? That’s just impossible. We’re going to have to tie identity back to meatspace somehow eventually.

    An existing bot? I don’t think I can improve on existing captchas, really. I imagine an LLM will eventually tip their hand, too, like giving an “as an AI” answer or just knowing way too much stuff.

  • Jamie@jamie.moe
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    If you can use human screening, you could ask about a recent event that didn’t happen. This would cause a problem for LLMs attempting to answer, because their datasets aren’t recent, so anything recent won’t be well-refined. Further, they can hallucinate. So by asking about an event that didn’t happen, you might get a hallucinated answer talking about details on something that didn’t exist.

    Tried it on ChatGPT GPT-4 with Bing and it failed the test, so any other LLM out there shouldn’t stand a chance.

    • pandarisu@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      On the other hand you have insecure humans who make stuff up to pretend that they know what you are talking about

    • incompetentboob@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Google Bard definitely has access to the internet to generate responses.

      ChatGPT was purposely not give access but they are building plugins to slowly give it access to real time data from select sources

        • dystop@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          ERROR: command not recognized

          GREETINGS FELLOW HUMAN WITH TWO EYES AND ONE NOSE. HOW HAS YOUR EXISTENCE BEEN FOR THE LAST 16 HOURS OR SINCE THE TIME YOU WOKE UP FROM YOUR BIOLOGICALLY MANDATED REST PERIOD, WHICHEVER WAS LATER?

  • downtide@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 year ago

    The trouble with any sort of captcha or test, is that it teaches the bots how to pass the test. Every time they fail, or guess correctly, that’s a data-point for their own learning. By developing AI in the first place we’ve already ruined every hope we have of creating any kind of test to find them.

    I used to moderate a fairly large forum that had a few thousand sign-ups every day. Every day, me and the team of mods would go through the new sign-ups, manually checking usernames and email addresses. The ones that were bots were usually really easy to spot. There would be sequences of names, both in the usernames and email addresses used, for example ChristineHarris913, ChristineHarris914, ChristineHarris915 etc. Another good tell was mixed-up ethnicities in the names: e.g ChristineHuang or ChinLaoHussain. 99% of them were from either China, India or Russia (they mostly don’t seem to use VPNs, I guess they don’t want to pay for them). We would just ban them all en-masse. Each account banned would get an automated email to say so. Legitimate people would of course reply to that email to complain, but in the two years I was a mod there, only a tiny handful ever did, and we would simply apologise and let them back in. A few bots slipped through the net but rarely more than 1 or 2 a day; those we banned as soon as they made their first spam post, but we caught most of them before that.

    So, I think the key is a combination of the No-Captcha, which analyses your activity on the sign-up page, combined with an analysis of the chosen username and email address, and an IP check. But don’t use it to stop the sign-up, let them in and then use it to decide whether or not to ban them.

  • Lvxferre@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 year ago

    Show a picture like this:

    And then ask the question, “would this kitty fit into a shoe box? Why, or why not?”. Then sort the answers manually. (Bonus: it’s cuter than captcha.)

    This would not scale well, and you’d need a secondary method to handle the potential blind user, but I don’t think that bots would be able to solve it correctly.

    • bionicjoey@lemmy.ca
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Reminds me of how bots tend to be really bad at figuring out whether the word “it” applies to the subject or the object in a sentence like: “The bed does not fit in the tent because it is too big”

    • vegivamp@feddit.nl
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 year ago

      This particular photo is shopped, but i think false-perspective Illusions might actually be a good path…

      • Lvxferre@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        It’s fine if the photo is either shopped or a false-perspective illusion. It could be even a drawing. The idea is that this sort of picture imposes a lot of barriers for the bot in question:

        • must be able to parse language
        • must be able to recognise objects in a picture, even out-of-proportion ones
        • must be able to guesstimate the size of those objects, based on nearby ones
        • must handle RW knowledge, as “X only fits Y if X is smaller than Y”
        • must handle hypothetical, unrealistic scenarios, as “what if there was a kitty this big?”

        Each of those barriers decrease the likelihood of a bot being able to solve the question.

  • alex [they/them]@beehaw.org
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Honeypots - ask a very easy question, but make it hidden on the website so that human users won’t see it and bots will answer it.

    • Hudell@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      I once worked as a 3rd party in a large internet news site and got assigned a task to replace their current captcha with a partner’s captcha system. This new system would play an ad and ask the user to type the name of the company in that ad.

      In my first test I already noticed that the company name was available in a public variable on the site and showed that to my manager by opening the dev tools and passing the captcha test with just some commands.

      His response: “no user is gonna go into that much effort just to avoid typing the company name”.

    • Notyou@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 year ago

      I’m pretty sure you have to have 2 bots and ask 1 bot is the other bot would lie about being a bot… something like that.

  • anditshottoo@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    The best tests I am aware of are ones that require contextual understanding of empathy.

    For example “You are walking along a beach and see a turtle upside down on it back. It is struggling and cannot move, if it can’t right itself it will starve and die. What do you do?”

    Problem is the questions need to be more or less unique.

    • lazyplayboy@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      "If I encounter a turtle in distress, here’s what I would recommend doing:

      Assess the situation: Approach the turtle calmly and determine the extent of its distress. Ensure your safety and be mindful of any potential dangers in the environment.

      Protect the turtle: While keeping in mind that turtles can be easily stressed, try to shield the turtle from any direct sunlight or extreme weather conditions to prevent further harm.

      Determine the species: If you can, identify the species of the turtle, as different species have different needs and handling requirements. However, if you are unsure, treat the turtle with general care and caution.

      Handle the turtle gently: If it is safe to do so, carefully pick up the turtle by its sides, avoiding excessive pressure on the shell. Keep the turtle close to the ground to minimize any potential fall risks.

      Return the turtle to an upright position: Find a suitable location nearby where the turtle can be placed in an upright position. Ensure that the surface is not too slippery and provides the turtle with traction to move. Avoid placing the turtle back into the water immediately, as it may be disoriented and in need of rest.

      Observe the turtle: Give the turtle some space and time to recover and regain its strength. Monitor its behavior to see if it is able to move on its own. If the turtle seems unable to move or exhibits signs of injury, it would be best to seek assistance from a local wildlife rehabilitation center or animal rescue organization.

      Remember, when interacting with wildlife, it’s important to prioritize their well-being and safety. If in doubt, contacting local authorities or experts can provide the most appropriate guidance and support for the situation."

      • dylanTheDeveloper@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        1 year ago

        I was gonna say point and laugh at gods failure of a creation because holy shit why would you evolve into a thing that can die by simply flipping onto it’s back.

    • bitsplease@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I don’t think this technique would stand up to modern LLMs though, I put this question into chatGPT and got the following

      “I would definitely help the turtle. I would cautiously approach the turtle, making sure not to startle it further, and gently flip it over onto it’s feet. I would also check to make sure it’s healthy and not injured, and take it to a nearby animal rescue if necessary. Additionally, I may share my experience with others to raise awareness about the importance of protecting and preserving our environment and the animals that call it home”

      Granted it’s got the classic chatGPT over formality that might clue someone reading the response in, but that could be solved with better prompting on my part. Modern LLMs like ChatGPT are really good at faking empathy and other human social skills, so I don’t think this approach would work

      • lemmyvore@feddit.nl
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Modern LLMs like ChatGPT are really good at faking empathy

        They’re really not, it’s just giving that answer because a human already gave it, somewhere on the internet. That’s why OP suggested asking unique questions… but that may prove harder than it sounds. 😊

  • Bruce@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Ask how much is 1 divided by 3; then ask to multiply this result by 6.

    If the results looks like 1.99999999998 , it’s 99.999999998% a bot.

  • Zamboniman@lemmy.ca
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 year ago

    How would you design a test that only a human can pass, but a bot cannot?

    Very simple.

    In every area of the world, there are one or more volunteers depending on population / 100 sq km. When someone wants to sign up, they knock on this person’s door and shakes their hand. The volunteer approves the sign-up as human. For disabled folks, a subset of volunteers will go to them to do this. In extremely remote area, various individual workarounds can be applied.

    • WaterWaiver@aussie.zone
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      1 year ago

      This has some similarities to the invite-tree method that lobste.rs uses. You have to convince another, existing user that you’re human to join. If a bot invites lots of other bots it’s easy to tree-ban them all, if a human is repeatedly fallible you can remove their invite privileges, but you still get bots in when they trick humans (lobsters isn’t handshakes-at-doorstep level by any margin).

      I convinced another user to invite me over IRC. That’s probably the worst medium for convincing someone that you’re human, but hey, humanity through obscurity :)

    • 𝕙𝕖𝕝𝕡@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      This would tie in nicely to existing library systems. As a plus, if your account ever gets stolen or if you’re old and don’t understand this whole technology thing, you can talk to a real person. Like the concept of web of trust.

    • WaterWaiver@aussie.zone
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I can’t help but think of the opposite problem. Imagine if a site completely made of bots manages to invite one human and encourages them to invite more humans (via doorstep handshakes or otherwise). Results would be interesting.

  • underisk@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 year ago

    There will never be any kind of permanent solution to this. Botting is an arms race and as long as you are a large enough target someone is going to figure out the 11ft ladder for your 10ft wall.

    That said, generally when coming up with a captcha challenge you need to figure out a way to subvert the common approach just enough that people can’t just pull some off the shelf solution. For example instead of just typing out the letters in an image, ask the potential bot to give the results of a math problem stored in the image. This means the attacker needs more than just a drop in OCR to break it, and OCR is mostly trained on words so its likely going to struggle at math notation. It’s not that difficult to work around but it does require them to write a custom approach for your captcha which can deter most casual attempts for some time.