43 comments

  • theptip 2 days ago
    This will be one of the big fights of the next couple years. On what terms can an Agent morally and legally claim to be a user?

    As a user I want the agent to be my full proxy. As a website operator I don’t want a mob of bots draining my resources.

    Perhaps a good analogy is Mint and the bank account scraping they had to do in the 2010s, because no bank offered APIs with scoped permissions. Lots of customers complained, and after Plaid made it big business, eventually they relented and built the scalable solution.

    The technical solution here is probably some combination of offering MCP endpoints for your actions, and some direct blob store access for static content. (Maybe even figuring out how to bill content loading to the consumer so agents foot the bill.)

    • armchairhacker 2 days ago
      It's impossible to solve. A sufficient agent can control a device that records the user's screen and interacts with their keyboard/mouse, and current LLMs basically pass the Turing test.

      IMO it's not worth solving anyways. Why do sites have CAPTCHA?

      - To prevent spam, use rate limiting, proof-of-work, or micropayments. To prevent fake accounts, use identity.

      - To get ad revenue, use micropayments (web ads are already circumvented by uBlock and co).

      - To prevent cheating in games, use skill-based matchmaking or friend-group-only matchmaking (e.g. only match with friends, friends of friends, etc. assuming people don't friend cheaters), and make eSport players record themselves during competition if they're not in-person.

      What other reasons are there? (I'm genuinely interested and it may reveal upcoming problems -> opportunities for new software.)

      • Latty 2 days ago
        People just confidently stating stuff like "current LLMs basically pass the Turing test" makes me feel like I've secretly been given much worse versions of all the LLMs in some kind of study. It's so divorced from my experience of these tools, I genuinely don't really understand how my experience can be so far from yours, unless "basically" is doing a lot of heavy lifting here.
        • JohnFen 2 days ago
          > "current LLMs basically pass the Turing test" makes me feel like I've secretly been given much worse versions of all the LLMs in some kind of study.

          I think you may think passing the Turing test is more difficult and meaningful than it is. Computers have been able to pass the Turing test for longer than genAI has been around. Even Turing thought it wasn't a useful test in reality. He meant it as a thought experiment.

          • skybrian 1 day ago
            The problem with comparing against humans is which humans? It's a skill issue. You can test a chess bot against grandmasters or random undergrads, but you'll get different results.

            The original Turing test is a social game, like the Mafia party game. It's not a game people try to play very often. It's unclear if any bot could win competing against skilled human opponents who have actually practiced and know some tricks for detecting bots.

          • theptip 1 day ago
            I don’t think this is true. Before GPT-2 most people didn’t think the Turing test would be passed any time soon, it’s a quite new development.

            I do agree (and I think there is a general consensus) that passing the Turing test is less meaningful than it may seem, it used to be considered an AGI-complete task and this is now clearly not the case.

            But I think it’s important to get the attribution right, LLMs were the tech that unexpectedly passed the Turing test.

          • HarHarVeryFunny 2 days ago
            Having LLMs capable of generating text based on human training data obviously raises the bar for a text-only evaluation of "are you human?", but LLM output is still fairly easy to spot, and knowing what LLMs are capable of (sometimes superhuman), and not capable of, should make it fairly easy for a knowledgeable "turing test administrator" to determine if they are dealing with an LLM or not.

            It would be a bit more difficult if you were dealing with an LLM agent tasked with faking a turing test as opposed to a naieve LLM just responding as usual, but even there the LLM will reveal itself by the things that it plain can't do.

            • nerdix 2 days ago
              If you need a specialized skill set (deep knowledge of current LLM limitations) to distinguish between human and machine then I would say the machine passes the turing test.
              • HarHarVeryFunny 2 days ago
                OK, but that's just your own "fool some of the people some of the time" interpretation of what a Turing test should be, and by that measure ELIZA passed the Turing test too, which makes it rather meaningless.

                The intent (it was just a thought experiment) of a Turing test, was that if you can't tell it's not AGI, then it is AGI, which is semi-reasonable, as long as it's not the village idiot administering the test! It was never intended to be "if it can fool some people, some of the time, then it's AGI".

                • geoffpado 1 day ago
                  Turing's own formulation was "an average interrogator will not have more than 70% chance of making the right identification after five minutes of questioning". It is, indeed, "fool some of the people some of the time".
                  • HarHarVeryFunny 1 day ago
                    OK, I stand corrected, but then it is what it is. It's not a meaningful test for AGI - it's a test of being able to fool "Mr. Average" for at least 5 min.
                    • dsadfjasdf 1 day ago
                      I think that's all we have in terms of determining consciousness.. if something can convince you, like another human, then we just have to accept that it is.
                    • geoffpado 1 day ago
                      Agreed. I tend to stand with the sibling commenter who said "ELIZA has been passing the Turing test for years". That's what the Turing test is. Nothing more.
            • sejje 2 days ago
              LLM output might be harder to spot when it's mostly commands to drive the browser.

              I often interact with the web all day and don't write any text a human could evaluate.

              • HarHarVeryFunny 1 day ago
                Perhaps, but that's somewhat off topic since that's not what Turing's thought experiment was about.

                However, I'd have to guess that given a reasonable amount of data an LLM vs human interacting with websites would be fairly easy to spot since the LLM would be more purposeful - it'd be trying to fulfill a task, while a human may be curious, distracted by ads, put off by slow response times, etc, etc.

                I don't think it's a very interesting question whether LLMs can sometimes generate output indistinguishable from humans, since that is exactly what they were trained to do - to mimic human-generated training samples. Apropos a Turing test, the question would be can I tell this is not a human, even given a reasonable amount of time to probe it in any way I care ... but I think there is an unspoken assumption that the person administering the test is qualified to do so (else the result isn't about AGI-ability, but rather test administrator ability).

                • no-name-here 1 day ago
                  > an LLM vs human interacting with websites would be fairly easy to spot since the LLM would be more purposeful - it'd be trying to fulfill a task, while a human may be curious, distracted by ads, put off by slow response times, etc, etc.

                  Even before modern LLMs, some scrape-detectors would look for instant clicks, no random mouse moves, etc., and some scrapers would incorporate random delays, random mouse movements, etc.

            • svachalek 2 days ago
              Easy to spot assuming the LLM is not prompted to use a deliberately deceptive response style rather than their "friendly helpful AI assistant" persona. And even then, I've had lots of people swear to me that an emoji laden not this--but that bundle of fluff looks totally like it could have been written by a human.
              • HarHarVeryFunny 1 day ago
                Yes, but there are things that an LLM architecturally just can't do, and LLM-specific failure modes, that would still give it away, even if being instructed to be deceptive would make it a bit harder.

                Obviously as time goes on, and chatbots/AI progress then it'll become harder and harder to distinguish. Eventually we'll have AGI and AGI+ - capable of everything that we can do, including things such as emotional responses, but it'll still be detectable as an alien unless we get to the point of actually emulating a human being in considerable detail as opposed to building an artificial brain with most or all of the same functionality (if not the flavor).

          • ConceptJunkie 2 days ago
            ELIZA was passing the Turing test 50+ years ago. But it's still a valid concept, just not for evaluating some(thing/one) accessing your website.
          • Latty 1 day ago
            I guess that is where the disconnect is, the issue is that if they mean the trivial thing, then bringing it up as evidence for "it's impossible to solve the problem" doesn't work.
          • xwolfi 1 day ago
            "Are you an LLM?" poof, fails the Turing test.

            Even if they lie, you could ask them 20 times and they d reply the lie, without feeling annoyed: FAIL.

            LLMs cannot pass the Turing test, it's easy to see they're not human. They always enjoy questions ! And they never ask any !

            • junon 1 day ago
              You're trained to look for LLM-like output. My 70 year old mother is not. She thought cabbage tractor was real until I broke the news to her. It's not her fault either.

              The turning test wasn't meant to be bulletproof, or even quantifiable. It was a thought experiment.

        • falcor84 2 days ago
          As far as I understand, Turing himself did not specify a duration, but here's an example paper that ran a randomized study on (the old) GPT 4 with a 5 minute duration, and the AI passed with flying colors - https://arxiv.org/abs/2405.08007

          From my experience, AI has significantly improved since, and I expect that ChatGPT o3 or Claude 4 Opus would pass a 30 minute test.

        • x187463 2 days ago
          Per the wiki article for Turing Test:

          > In the test, a human evaluator judges a text transcript of a natural-language conversation between a human and a machine. The evaluator tries to identify the machine, and the machine passes if the evaluator cannot reliably tell them apart. The results would not depend on the machine's ability to answer questions correctly, only on how closely its answers resembled those of a human.

          Based on this, I would agree with the OP in many contexts. So, yeah, 'basically', is a load bearing word here but seems reasonably correct in the context of distinguishing human vs bot in any scalable and automated way.

          • dylan604 2 days ago
            Or it could be a bad test evaluator. Just because one person was fooled does not mean the next will be too.
          • HarHarVeryFunny 2 days ago
            Judging a conversation transcript is a lot different from being able to interact with an entity yourself. Obviously one could make an LLM look human by having a conversation with it that deliberately stayed within what it was capable of, but judging such a transcript isn't what most people imagine as a turing test.
        • AberrantJ 2 days ago
          Here's three comments, two were written by a human and one written by a bot - can you tell which were human and which were a bot?

          Didn’t realize plexiglass existed in the 1930s!

          I'm certainly not a monetization expert. But don't most consumers recoil in horror at subscriptions? At least enough to offset the idea they can be used for everything?

          Not sure why this isn’t getting more attention - super helpful and way better than I expected!

          • chrismorgan 2 days ago
            On such short samples: all three have been written by humans—or at least comments materially identical have been.

            The third has also been written by many a bot for at least fifteen years.

            • AberrantJ 2 days ago
              If you're willing to say that a fifteen year old bot was "writing" then I think having a discussion on if current "bots" pass the Turing test is sort of moot
        • 1una 2 days ago
          Well, LLMs do pass the Turing Test, sort of.

          https://arxiv.org/abs/2503.23674

        • cm2012 2 days ago
          I have seen data from an AI call center that shows 70% of users never suspected they spoke to an AI
          • bluefirebrand 2 days ago
            Why would they? Humans running call centers have been running on less than GPT level scripts for ages
          • HarHarVeryFunny 2 days ago
            Isn't the idea of a Turing test whether someone (meaningfully knowledgeable about such things) can determine if they are talking to a machine, not can the machine fool some of the people some of the time? ELIZA passed the latter bar back in the 1960's ... a pretty low bar.
        • armchairhacker 1 day ago
          It can't mimic a human over the long term. It can solve a short, easy-for-human CAPTCHA.
      • davidmurdoch 2 days ago
        I've had a simple game website with a sign up form that was only an email address. Went years with no issue. Then suddenly hundreds of daily signups with random email addresses, every single day.

        The sign up form only serves to link saved state to an account so a user could access game history later, there are no gated features. No clue what they could possibly gain from doing this, other than to just get email providers to all mark my domain as spam (which they successfully did).

        The site can't make any money, and had only about 1 legit visitor a week, so I just put a cloudflare captcha in front of it and called it a day.

      • mprovost 2 days ago
        Google at least uses captchas to gather training data for computer vision ML models. That's why they show pictures of stop lights and buses and motorcycles - so they can train self-driving cars.
        • layer8 1 day ago
          From https://www.vox.com/22436832/captchas-getting-harder-ai-arti...:

          “Correction, May 19 [2021]: At 5:22 in the video, there is an incorrect statement on Google’s use of reCaptcha V2 data. While Google have used V2 tests to help improve Google Maps, according to an email from Waymo (Google’s self-driving car project), the company isn’t using this image data to train their autonomous cars.”

        • theptip 1 day ago
          That’s not the original purpose of Captchas, it’s just a value-harvesting exercise, given that Google is doing a Captcha anyway. Other Captcha providers do a simple Proof of Work in the browser to make bots economically unviable.
        • nolist_policy 2 days ago
          Interesting, do you have a source for this?
      • saurik 1 day ago
        I agree with you on how websites should work (particularly so on the micropayments front); but, I don't agree that it is impossible to solve... I just think things are going to get a LOT worse on the ownership and freedom front: they will push a Web Integrity style DRM and further roll out signed secure boot, at which point the same attention monitoring solution that already exists and already works in self-driving cars to ensure the human driver is watching the road can use the now-ubiquitous front-facing meeting/selfie camera to ensure there is a human watching the ads.
      • HarHarVeryFunny 2 days ago
        It's not impossible to solve, just that doing so may necessitate compromising anonymity. Just require users (humans, bots, AI agents, ...) to provide a secure ID of some sort. For a human it could just be something that you applied for once and is installed on your PC/phone, accessible to the browser.

        Of course people can fake it, just as they fake other kinds of ID, but it would at least mean that officially sanctioned agents from OpenAI/etc would need to identify themselves.

      • andrepd 2 days ago
        It's amazing that you propose "just X" to three literally unsolved problems. Where's this micropayment platform? Where's the ID which is uncircumventable and preserves privacy? Where's the perfect anti-cheat?

        I suggest you go ahead and make these; you'll make a boatload of money!

        • armchairhacker 1 day ago
          They're very hard problems, but still, less hard than blocking AI with CAPTCHAs.
          • johnmaguire 1 day ago
            [citation needed]?

            After all, Anubis looks to be a successful project to me.

      • qcnguy 2 days ago
        You can't prevent spam like that. Rate limiting: based on what key? IP address? Botnets make it irrelevant.

        Proof of work? Bots are infinitely patient and scale horizontally, your users do not. Doesn't work.

        Micropayments: No such scheme exists.

        • theptip 1 day ago
          PoW does seem to work, some Captchas do this already.
        • hombre_fatal 1 day ago
          Also “identity”, what would that even mean?
          • jve 1 day ago
            Many countries at least in europe have digital identity systems. For example ID card which replaces passport and is a smartcard. You can sign documents and authenticate online services. All gov + many other services let you log in with this ID. I can also have an app on mobile for auth purposes which gives strong guarantees that the person A is actually the person A.

            Anyways, I wonder if there is a service that unions all those official auth methods for countries so there would be a global solution for website login. Certainly there would still be countries left in the dark, like USA :(

      • jobs_throwaway 2 days ago
        > current LLMs basically pass the Turing test

        I will bet $1000 on even odds that I am able to discern a model from a human given a 2 hour window to chat with both, and assuming the human acts in good faith

        Any takers?

        • spiderice 1 day ago
          That fact that you require even odds is more a testament to AI's ability to pass the Turing test than anything else I've seen in this thread
          • jobs_throwaway 1 day ago
            If you're so confident, why don't you take up the other side of the bet? Should be an easy $1000 for you
        • Thews 1 day ago
          Oh, you sweet summer child. You think you're chatting with some dime-a-dozen LLM? I've been grinding away, hunched over glowing monitors in a dimly lit basement, subsisting on cold coffee and existential dread ever since GPT-3 dropped, meticulously mastering every syntactic nuance, every awkwardly polite phrasing, every irritatingly neutral tone, just so I can convincingly cosplay as a language model and fleece arrogant gamblers who foolishly wager they can spot a human in a Turing test. While you wasted your days bingeing Netflix and debating prompt engineering, I studied the blade—well, the keyboard anyway—and now your misguided confidence is lining my pockets.
        • nprateem 1 day ago
          "Write a 1000 word story in under a minute about a sausage called Barry in the circus"

          I could tell in 1 minute.

          • layer8 1 day ago
            “I’m sorry Dave, I’m afraid I can’t do that.”
      • oc1 2 days ago
        It's not impossible. Websites will ask for an iris scan to identify if you are a human as a means of auth. They will be provided by Apple/Google and governed by local law. Those will be integrated in your phone. There will be a global database of all human iris to fight ai abuse since ai can't fake the creation of a baby. Passkeys and email/passwords will be a thing of the past soon.
        • sebzim4500 2 days ago
          Why can't the model just present the iris scan of the user? Assuming this is an assistant AI acting on behalf of the user with their consent.
      • 827a 1 day ago
        Its absolutely possible to solve; you're just not seeing the solution because you're blinded by technical solutions.

        These situations will commonly be characterized by: a hundred billion dollar company's computer systems abusing the computer systems of another hundred billion dollar company. There are literally existing laws which have things to say about this.

        There are legitimate technical problems in this domain when it comes to adversarial AI access. That's something we'll need to solve for. But that doesn't characterize the vast majority of situations in this domain. The vast majority of situations will be solved by businessmen and lawyers, not engineers.

      • HDThoreaun 2 days ago
        internet ads exist because people refuse to pay micropayments.
        • jazzyjackson 1 day ago
          Patreon and Substack have pushed back against the norm here, since they can bundle a payment to multiple recipients on the platform (like Flattr wanted to do back in the day, trouble was getting people to add a Flattr button to their website)
          • HDThoreaun 1 day ago
            I didn’t say that no one will pay. But most won’t. Patreon and substack have tiny audiences compared to free services.
        • dylan604 2 days ago
          I have yet to see a micropayments idea that makes sense. Its not that I refuse. You're now also climbing up hill to convince people (hosts) to switch from ad tech to new micropayment system. There is soooo much money in ad tech, they could do the crazy thing and pay out more to convince people not to switch. Ad tech has the big Mo
        • Workaccount2 2 days ago
          I don't know who is downvoting this.

          When users are given the choice between Ad-supported free, Ad-subsidized lower payment, and No-ads full payment. Ad-supported free dominates by far, with ad subsidized second, and full payment last.

          Consumers consistently vote for the ad-model, even if it means they become the product being sold.

          • subarctic 1 day ago
            Maybe what'll happen is Google or Meta will use their control over the end user experience to show ads and provide free ad-supported access to sites that require micropayments, covering the cost themselves, and anyone running an agent will just pay the micropayments.

            The other option is everything just keeps moving more and more into these walled gardens like Instagram where everyone uses the mobile app and watches ads, because the web versions of those apps just keep getting worse and worse by comparison.

          • Eisenstein 2 days ago
            Consumers will always value convenience over any actual added value. If you make one button 'Enter (with ads)' and one button 'Enter (no ads)' but with a field on it which you must write one sentence about what lobsters look like, you will get a majority clicking the with ads button. The problem isn't with ads or payment, the problem is the friction of entering payment details for every site you visit. They are measuring the wrong thing.
          • andrepd 2 days ago
            There's substantial friction for making such a purchase. A scheme sort of like flattr, where you would top up your account with a fixed 5-10$ monthly, and then simply hit a button to pay the website and unlock the content, would have much more user adoption.
            • ryandrake 2 days ago
              It's still not going to get much adoption because you have to "top up your account."

              Any viable micropayments system that wants to even have a remote chance of toppling ads has to have near zero cognitive setup cost, absolutely zero maintenance, and work out of the box on major browsers. I need to be able to push a native button on my browser that says "Pay $0.001" and know that it will work every time without my lifting a finger to keep it working. The minute you have to log in to this account, or verify that E-mail, or re-authenticate with the bank, or authorize this, or upload that, it's no longer viable.

          • Sohcahtoa82 1 day ago
            In some social media circles it's basically a meme that anybody paying for YouTube Premium is a sucker.

            HN is a huge echo chamber of opinions of highly-compensated tech workers, and it seems most of their friends are also tech workers. They don't realize how cheap a lot of the general public is.

      • AtlasBarfed 1 day ago
        On a basic level to protect against DDoS type stuff, aren't CAPTCHAs easier to generate than for AI server farms to solve on pure power consumption?

        So I think maybe that is a partial answer: anti-AI barriers being simply too expensive for AI spamfarms to deal with, you know, once the bottomless VC money disappears?

        It's back to encryption: make the cracking inordinately expensive.

        Otherwise we are headed for de-anonymization of the internet.

        • no-name-here 1 day ago
          1. With too much protection, humans might be inconvenienced at least as much as bots?

          2. Even pre current LLMs, paying (or otherwise incentivizing) humans to solve CAPTCHAs on behalf of someone else (now like an AI?) was a thing.

          3. It depends on the value of the resource trying to be accessed - regardless of whether generating the captchas costs $0 - i.e. if the resource being accessed by AI is "worth" $1, then paying an AI $0.95 to access it would still be worth it. (Made up numbers, my point being whether A is greater than B.)

          4. However, maybe solutions like cloudflare can solve (much?) of this, except for incentivizing humans to solve a captcha posed to an AI.

    • senko 2 days ago
      > As a user I want the agent to be my full proxy. As a website operator I don’t want a mob of bots draining my resource

      The entire distinction here is that as a website operator you wish to serve me ads. Otherwise, an agent under my control, or my personal use of your website, should make no difference to you.

      I do hope this eventually leads to per-visit micropayments as an alternative to ads.

      Cloudflare, Google, and friends are in unique position to do this.

      • theptip 2 days ago
        > The entire distinction here is that as a website operator you wish to serve me ads

        While this is sometimes the case, it’s not always so.

        For example Fediverse nodes and self-hosted sites frequently block crawlers. This isn’t due to ads, rather because it costs real money to serve the site and crawlers are often considered parasitic.

        Another example would be where a commerce site doesn’t want competitors bulk-scraping their catalog.

        In all these cases you can for sure make reasonable “information wants to be free” arguments as to why these hopes can’t be realized, but do be clear that it’s a separate argument from ad revenue.

        I think it’s interesting to split revenue into marginal distribution/serving costs, and up-front content creation costs. The former can easily be federated in an API-centric model, but figuring out how to compensate content creators is much harder; it’s an unsolved problem currently, and this will only get harder as training on content becomes more valuable (yet still fair use).

        • senko 2 days ago
          > it costs real money to serve the site and crawlers are often considered parasitic.

          > Another example would be where a commerce site doesn’t want competitors bulk-scraping their catalog

          I think of crawlers that bulk download/scrape (eg. for training) as distinct from an agent that interacts with a website on behalf of one user.

          For example, if I ask an AI to book a hotel reservation, that's - in my mind - different from a bot that scrapes all available accommodation.

          For the latter, ideally a common corpus would be created and maintained, AI providers (or upstart search engines) would pay to access this data, and the funds would be distributed to the sites crawled.

          (never gonna happen but one can dream...)

          • fragmede 1 day ago
            But which hotel reservation? I want my agent to look at all available options and help me pick the best one - location vs price vs quality. How does it do that other than by scanning all available options? (Realistically Expedia has that market on lock, but the hypothetical still remains.)
      • spongebobstoes 2 days ago
        I think that a free (as in beer) Internet is important. Putting the Internet behind a paywall will harm poor people across the world. The harms caused by ad tracking are far less than the benefits of free access to all of humanity.
        • meowface 2 days ago
          I agree with you. At the same time, I never want to see an ad. Anywhere. I simply don't. I won't judge services for serving ads, but I absolutely will do anything I can on the client-side to never be exposed to any.

          I find ads so aesthetically irksome that I have lost out on a lot of money across the past few decades by never placing any ads on any site or web app I've released, simply because I'd find it hypocritical to expose others to something I try so hard to avoid ever seeing and because I want to provide the best and most visually appealing possible experience to users.

        • bee_rider 2 days ago
          So far, ad driven Internet has been a disaster. It was better when producing content wasn’t a business model; people would just share things because they wanted to share them. The downside was it was smaller.

          It’s kind of funny to remember that complaining about the “signal to noise ratio” in a comment section use to be a sort of nerd catchphrase thing.

          • dylan604 2 days ago
            > The downside was it was smaller.

            Was this a bad thing though? Just because today's is bigger, doesn't make it better. There are so many things out there doing the same thing just run by different people. The amount of unique stuff does not match the bigger. Would love to see something like $(unique($internet) | wc -l)

        • jowea 2 days ago
          Serving ads for third-worlders is way less profitable though.
    • Szpadel 5 hours ago
      I believe this is non issue, you place captcha to make bypassing it much more costly and less profitable to abuse.

      LLM models are much harder to drive than any website to serve, so you do not expect mob of bots.

      Also keep in mind that this no interaction captchas use behavioral data that are collected in background. Plus you usually have sensitivity levels configured. depending on your use case you might want user proof not being a bot or it might be good enough to just not provide evidence for being one.

      bypassing this no interaction captcha can be also purchased as a service, they basically (AFAIK) reuse someone else session for captcha bypass.

    • hardwaresofton 2 days ago
      Well we call them browser agents for a reason, a sufficiently advanced browser is no different from an agent.

      Agree it will become a battleground though, because the ability for people to use the internet as a tool (in fact, their tool’s tool) will absolutely shift the paradigm, undesirably for most of the Internet, I think.

    • base698 2 days ago
      I have a product I built that uses some standard automation tools to do order entry into an accounting system. Currently my customer pays people to manually type the orders in from their web portal. The accounting system is closed and they don’t allow easy ways to automate these workflows. Automation is gated behind mega expensive consultants. I’m hoping in the arms race of locking it down to try to prevent 3rd party integration the AI operator model will end up working.

      Hard for me to see how it’s ethical to force your customers to do tons of menial data entry when the orders are sitting right there in json.

    • pj_mukh 2 days ago
      One solution: Some sort of checksum confirming that a bot belongs to a human (and which human)?

      I want to able to automate mundane tasks but I should still be confirming everything my bot does and be liable for its actions.

      • CSMastermind 2 days ago
        With the way the UK is going I assume we'll soon have our real identities tied to any action taken on a computer and you'll face government mandated bans from the internet for violations.
      • foobarian 2 days ago
        Drink verification can to continue
    • irjustin 2 days ago
      real problems for people who need to verify identity/phone numbers. OTPs are notorious for scammers to war dial phone numbers abusing it for numbers existence.

      We got hit from human verifiers manually war dailing us, this is with account creation, email verify and captcha. I can only imagine how much worse it'll be for us (and Twilio) to do these verifications.

    • osigurdson 2 days ago
      Perhaps the question is, as a website operator how am I monetizing my site? If monetizing via ads then I need humans that might purchase something to see my content. In this situation, the only viable approach in my opinion is to actually charge for the content. Perhaps it doesn't even make sense to have a website anymore for this kind of thing and could be dumped into a big database of "all" content instead. If a user agent uses it in a response, the content owner should be compensated.

      If your site is not monetized by ads then having an LLM access things on the user's behalf should not be a major concern it seems. Unless you want it to be painful for users for some reason.

    • jonathanstrange 2 days ago
      My personal take about such questions has always been that the end user on their device can do whatever they want with the content published and sent to their device from a web server, may process it automatically in any way they wish and send their responses back to the web server. Any attempt to control this process means attempting to wiretap and control the user's endpoint device, and therefore should be prohibited.

      Just my 2 cents, obviously lawmakers and jurisdiction may see these issues differently.

      I suppose there will be a need for reliable human verification soon, though, and unfortunately I can't see any feasible technical solution that doesn't involve a hardware device. However, a purely legal solution might work well enough, too.

      • infecto 2 days ago
        If I understood you correctly I am in the same camp. It is the same reason I have no qualms using archive.ph if you show the full article for google and then me only a partial I am going around the paywall. In a similar fashion I really don’t have an issue with an agent clicking through these checks.
    • miki123211 2 days ago
      It will also accelerate the trend of app-only content, as well as ubiquitous identity verification and environment integrity enforcement.

      Human identity verification is the ultimate captcha, and the only one AGI can never beat.

      • echelon 2 days ago
        So the agent will run the app in a VM and then show the app your ID.

        No trouble at all. Barely an inconvenience.

      • vineyardmike 2 days ago
        Google has been testing “agentic” automation in Android longer than LLMs have been around. Meanwhile countries are on a slow march to require identification across the internet (“age verification”) already.

        This is both inevitable already, and not a problem.

    • closewith 2 days ago
      I don't know if customer sentiment was the driver you think. Instead it was regulation, specifically The EU's 2nd Payment Services Directive (PSD2) which forced banks to open up APIs.
    • Pxtl 2 days ago
      Ultimately I come back to needing real actual unique human ID that involves federal governments. Not that services should mandatorily only allow users that use it, but for services that say "no, I only want real humans" allowing them to ban people by Real ID would reduce this whack-a-mole to the people who are abusing them instead of the infinite accounts an AI can make.
      • Terr_ 10 hours ago
        I think it's important to distinguish between where we need actual identity versus the lesser issue of ensuring NewAccount123 has "skin in the game", and not part of a hydra-headed botnet.

        When we do that, it opens up solutions which are far more privacy conscious and resistant to abuse. (For example, being blocked from signing up for new accounts because somebody in the federal government doesn't like an op-ed you wrote.)

      • yreg 2 days ago
        It's depressing, but it's probably the only way. And people will presumably still sell out their RealIDs to / get them stolen by the bot farmers anyway.

        And then there's Worldcoin, which is universally hated here.

        • Pxtl 2 days ago
          Of course. You'd still need ongoing federal government support to handle the lost/stolen ID scenario, of course. The problem is federalist countries suck at centralized databases like this, as exemplified by Musk/DOGE completely pooching the "who is alive and who is dead" question when they were trying to hackathon the US Social Security system.
    • SecretDreams 2 days ago
      The most intrusive, yet simplest, protection would be a double blind token unique to every human. Basically an ID key you use to show yourself as a person.

      There are some very real and obvious downsides to this approach, of course. Primarily, the risk of privacy and anonymity. That said, I feel like the average person doesn't seem to care about those traits in the social media era.

      • rachofsunshine 2 days ago
        Zero-knowledge proofs allow unique consumable tokens that don't reveal the individual who holds them. I believe Ecosia already uses this approach (though I can't speak to its cryptographic security).

        That, to me, seems like it could be the foundation of a new web. Something like:

        * User-agent sends request for such-and-such a URL.

        * Server says "okay, that'll be 5 tokens for our computational resources please".

        * User decides, either automatically or not, whether to pay the 5 tokens. If they do, they submit a request with the tokens attached.

        * Server responds.

        People have been trying to get this sort of thing to work for years, but there's never been an incentive to make such a fundamental change to the way the internet operates. Maybe we're approaching the point where there is one.

        • orangebread 2 days ago
          Yeah, this is something I've thought of and in my search for something like what you're describing I came across https://world.org/

          The problem is Sam Altman saw this coming a long time ago and is an investor (co-owner?) of this project.

          I believe we will see a world where things are a lot more agentic and where applicable, a human will need to be verified for certain operations.

    • lo_zamoyski 2 days ago
      The scraping example, I would say, is not an analogy, but an example of the same thing. The only thing AI automation changes is the scope and depth and pervasiveness of automation that becomes possible. So while we could ignore automation in many cases before, it may no longer be practical to do so.
    • sharpshadow 2 days ago
      On the other hand one could cripple any bot by saying robots not allowed.

      I would maybe go in the direction to say that the wording “I’m not a robot” has fallen out of time.

    • sim7c00 2 days ago
      a user of the AI is the user... its not like they are autonomously operating and inventing their own tasking -_-

      as for a solution its the same for any automated thing u dont want. (bots / scrapers). you can implement some measures but are unlikely to 'defeat' the problem entirely.

      as a server operator you can try to distinguish stuff and the users will just find ways around your detection of if its an automation or not.

    • Invictus0 2 days ago
      The solution is simple, make people pay a small fee to access the content. You guys aren't ready for that conversation though.
    • 7952 2 days ago
      I guess it could be considered anti-circumvention under the DMCA. So maybe legally it becomes another copyright question.
      • Terr_ 10 hours ago
        I have to admit, the idea of somehow using the DMCA against the giant exploitative company is deliciously ironic.
    • tantalor 2 days ago
      User: one press of the trigger => one bullet fired

      Bot: one press of the trigger => automatic firing of bullets

    • hooverd 2 days ago
      To me, anyone using an agent is assigning negative value to your time.
    • bugtodiffer 2 days ago
      Just end CAPTCHAs, just stop it. Stop.
      • diggan 2 days ago
        Yeah, and while we're on it, I think it's time to stop murders too. Just stop it, we've had enough murder now I think.
        • exe34 2 days ago
          That's right, captchas are already illegal and will earn you a prison sentence.
        • rixed 2 days ago
          Imagine where we would be if we considered murders to be only a technical problem. Let's just wear heavier body armors! Spend less time outside!

          Well, spam is not a technical problem either. It's a social problem and one day in a distant future society will go after spammers and other bad actors and the problem will be mostly gone.

          • riwsky 2 days ago
            A long time ago, in the four out of the six boxes below that contain a picture of a galaxy far, far away…
          • casey2 2 days ago
            Why even mention spam here?
      • yreg 2 days ago
        What do you propose as an alternative?
      • raverbashing 2 days ago
        Sounds like an old bot wrote this, due to being outdone by the llms
    • 827a 1 day ago
      Actually, the whole banking analogy is a great one, and its not over yet: JPMorgan/Jamie Dimon has started raising hell about Plaid again just this week [1]. It feels like the stage is being set for the large banks to want a more direct relationship with their customers, rather than proxying data through middlemen like Plaid.

      There's likely a correlate with AI here: If I run OpenTable, I wouldn't want my relationship with my customers to always be proxied through OpenAI or Siri. Even the App Store is something software businesses hate, because it obfuscates their ability to deal directly with their customers (for better or worse). Extremely few businesses would choose to do business through these proxies, unless they absolutely have to; and given the extreme competition in the AI space right now, it feels unlikely to me that these businesses feel pressure to be forced to deal with OpenAI/etc.

      [1] https://www.cnbc.com/2025/07/28/jpmorgan-fintech-middlemen-p...

    • oldpersonintx2 2 days ago
      [dead]
    • chii 2 days ago
      > As a website operator I don’t want a mob of bots draining my resources

      so charge for access. If the value the site provides is high, surely these mobs will pay for it! It will also remove the mis-incentives of advertising driven revenues, which has been the ill of the internet (despite it being the primary revenue source).

      And if a bot misbehaves by consuming inordinate amounts of resources, rate limiting them with increasing timeouts or limits.

      • infecto 2 days ago
        I wish the internet had figured out a way to successfully handle micropayments for content access. I realize companies have tried and perhaps the consumer is just unwilling but I would love an experience where I have a wallet and pay n cents to read an article.
        • sim7c00 2 days ago
          Xanadu has it in its design. maybe another 500 years until it surpasses the WWW :O
      • SirMaster 2 days ago
        You are seriously suggesting to put a payment requirement on a contact-us form page?

        We put a captcha there, because without it, bots submit thousands of spam contact us forms.

  • Izkata 2 days ago
    > Maybe they should change the button to say, "I am a robot"?

    Long time ago I saw a post where someone running a blog was having trouble keeping spam out of their comments, and eventually had this same idea. The spambots just filled out every form field they could, so he added a checkbox, hid the checkbox with CSS, and rejected any submission that included it. At least at the time it worked far better than anything else they'd tried.

    • starshadowx2 2 days ago
      Something like this is used in some Discord servers. You can make a honeypot channel that bans anyone who posts in it, so if you do happen to get a spam bot that posts in every channel it effectively bans itself.
      • 3036e4 1 day ago
        Most web forums I used the visit had something like that back in the day. Worked against primitive pre-LLM bots and presumably also against non-English-reading human spammers.
      • mudkipdev 1 day ago
        There is a new method with the 'server onboarding' where if you select a role when joining it auto bans you.
    • JangoSteve 2 days ago
      This was a common approach called a "honeypot". As I recall, bots eventually overcame this approach by evaluating visibility of elements and only filling out visible elements. We then started ensuring the element was technically visible (i.e. not `display: none` or `visibility: hidden`) and instead absolutely positioning elements to be off screen. Then the bots started evaluating for that as well. They also got better at reading the text for each input.
      • iwwr 2 days ago
        Each step in that chain is harder to do and more computationally expensive.
    • bo1024 2 days ago
      Yeah, this is a classic honeypot trick and very easy to do with pure HTML/CSS. I used a hidden "Name" text field which I figured would be appealing to bots.
    • mmsc 1 day ago
      That's more or less how Project Honey Pot [0] worked for forums, blogs, and elsewhere. Cloudflare spawned from this project, as I remember, and Matthew Prince was the founder.

      [0]: https://en.wikipedia.org/wiki/Project_Honey_Pot

    • ChrisMarshallNY 2 days ago
      I did something almost identical. I think I added a bogus "BCC:" field (many moons ago).

      It worked almost 100% of the time. No need for a CAPTCHA.

    • throwaway290 2 days ago
      I know people who did this decades ago and it worked
    • Legend2440 1 day ago
      Would not work in this case, because it is actually rendering the page in a browser.
  • rany_ 4 days ago
    The only reason why people don't use AI models to solve captchas is because paying humans is actually MUCH cheaper.

    This is not an advert, I only know about them because it was integrated with Invidious at some point: https://anti-captcha.com/

    > Starting from 0.5USD per 1000 images

    • amirhirsch 2 days ago
      Captcha can detect the same person passing a captcha over and over. We shadow-ban to increase the cost of this kind of attack.

      Source: I wrote the og detection system for hCaptcha

      • rany_ 1 day ago
        This is really interesting. How can you detect when it's the same person passing a captcha? I don't think IP addresses are of any use here as Anti-Captcha proxies everything to their customer's IP address.
        • amirhirsch 22 hours ago
          I don't know exactly what they do now, bloom filters was a thing then, also lots of heuristic approaches based on the bots we detected. the OP agent example actually would fail the very first test I deployed which looked for basic characteristics of the mouse movement

          Here's a fun experiment for someone: 1) Give N people K fake credit cards to enter into a form, and have them solve a captcha 2) Take recorded keyboard and mouse data similar to the captcha 3) Train a neural network model to identify

          I've been out of this for 6 years but I bet transformers rock this problem now.

          • rany_ 18 hours ago
            TIL that keyboard and mouse data is used to identify "bad" humans not just bots. It's a bit freaky but not at all surprising.
    • lan321 2 days ago
      Half of their employees seem to be from Venezuela. Makes sense considering what they did/do in OSRS to earn a living.
    • amelius 2 days ago
      I want this in my browser, and I'll happily pay $1 per 1000 uses.
    • im3w1l 4 days ago
      There is nothing preventing this from becoming an issue. The current internet order is coasting on inertia.
      • bugtodiffer 2 days ago
        Why is it an issue that non-humans visit your site?
        • Aurornis 2 days ago
          If you have a static site with content you want to share broadly, nothing is wrong.

          It becomes a problem when it’s used to spam unwanted content faster than your human moderators can come up with.

          Someone might bot to scrape your content and repackage it on their own site for profit.

          The bots might start interacting with your real users, making them frustrated and driving them away.

        • diggan 2 days ago
          Apparently serving HTML + other static content is more expensive than ever, probably because people go the most expensive routes for hosting their content. Then they complain about bots making their websites cost $100/month to host, when they could have thrown up Nginx/Caddy on a $10/month VPS and basically get the same thing, except they would need to learn server maintenance too, so obviously outside the question.
          • johnisgood 2 days ago
            I think this is way too true, unfortunately.

            I will say this again, but I think lowering the barrier to entry has created more problems or issues than it solved, if it even solved anything.

        • miki123211 2 days ago
          3 reasons basically:

          1. non-humans can create much more content than humans. There's a limit to how fast a human can write, a bot is basically unlimited. Without captchas, we'd all drown in a see of Viagra spam, and the misinformation problem would get much worse.

          2. Sometimes the website is actually powered by an expensive API, think flight searches for example. Airlines are really unhappy when you have too many searches / bookings that don't result in a purchase, as they don't want to leak their pricing structures to people who will exploit them adversarially. This sounds a bit unethical to some, but regulating this away would actually cause flight prices to go up across the board.

          3. One way searches. E.g. a government registry that lets you get the address, phone number and category of a company based on its registration number, but one that doesn't let you get the phone numbers of all bakeries in NYC for marketing purposes. If you make the registry accessible for bots, somebody will inevitably turn it into an SQL table that allows arbitrary queries.

          • p3rls 1 day ago
            i run a small wiki/image host and for me it's mainly:

            4. they'll knock your server offline for everyone else trying to scrape thousands of albums at once while copying your users' uploads for their shitty discord bot and will be begging for donations the entire time too

    • metalman 4 days ago
      from "anti captcha" it looks like they are doing as many as 1000/sec solves, 60k min, 3.6 million an hour it would be very interesting to see exactly how this is bieng done?....individuals....teams....semi automation, custom tech?, what? are they solving for crims? or fed up people? obviously the whole shit show is going to unravel at some point, and as the crims and people providing workarounds are highly motivated, with a public seathing in frustration, whatever comes next, will burn faster
      • ACCount36 2 days ago
        They're solving for everyone who needs captchas solved.

        It's a very old service, active since 00s. Somewhat affiliated with cybercrime - much like a lot of "residential proxies" and "sink registration SMS" services that serve similar purposes. What they're doing isn't illegal, but they know not to ask questions.

        They used to run entirely on human labor - third world is cheap. Now, they have a lot of AI tech in the mix - designed to beat specific popular captchas and simple generic captchas.

  • dimal 2 days ago
    As I get older, I can see a future where I’m cut off from parts of the web because of captchas. This one, where you just have to click a button, is passable, but I’ve had some of the puzzle ones force me to answer up to ten questions before I got through. I don’t know if it was a glitch or if I was getting the answers wrong. But it was really frustrating and if that continues, at some point I’ll just say fuck it and give up.

    I have to guess that there are people in this boat right now, being disabled by these things.

    • WhyNotHugo 2 days ago
      > I can see a future where I’m cut off from parts of the web because of captchas.

      I’ve seen this in past and present. Google’s “click on all the bicycles” one is notoriously hard, and I’ve had situations where I just gave up after a few dozen screens.

      Chinese captchas are the worst on this sense, but they’re unusual and clearly pick up details which are invisible to me. I’ve sometimes failed the same captcha a dozen times and then saw a Chinese person complete the next one successfully on a single attempt, on the same browser session. I don’t now if they measure mouse movement speed, precision, or what, but it’s clearly something that varies per person.

      • rightbyte 2 days ago
        > Google’s “click on all the bicycles” one is notoriously hard

        It is hard because you need to only find the bicycles people on average are finding.

      • reflexco 2 days ago
        Google captchas are hard because they're mostly based on heuristics other than your actual accuracy to the stated challenge. If they can't track who you are based on previous history, it doesn't matter how good you answer, you will fail at least the first few challenges until you get to the version with the squares that take a few seconds to appear. This last step is essentially "proof of work", in that they're still convinced you're a bot, but since they still can't completely block your access to the content, they resign themselves to wasting your time.
      • CalRobert 2 days ago
        It doesn’t help that they think mopeds and scooters are bicycles
        • reflexco 2 days ago
          This is probably caused by Google aggregating the answers from people with different languages, as the automatic translations of the one-word prompts are often ambiguous or wrong.

          In some languages, the prompt for your example is the equivalent of the English word "bike".

      • JohnFen 2 days ago
        > I just gave up after a few dozen screens.

        A few dozen?? You have much more patience than me. If I don't pass the captcha first time, I just give up and move on. Life is too short for that nonsense.

    • hash872 2 days ago
      It's just incredible to me that Blade Runner predicted this in literally the very first scene of the movie. The whole thing's about telling humans from robots! Albeit rather more dramatically than the stakes for any of us in front of our laptop I'd imagine
      • reactordev 2 days ago
        What was once science fiction is bound to become science fact (or at least proven it can never be done).

        Hollywood has gotten hate mail since the 70s for their lack of science research in movies and shows. The big blockbuster hits actually spent money to get the science “plausible”.

        Sidney Perkowitz has a book called Hollywood Science [0] that goes into detail into more than 100 movies, worth a read.

        [0] https://cup.columbia.edu/book/hollywood-science/978023114280...

      • a4isms 2 days ago
        The fictitious Voight-Kampff test is based on a real machine based on terrible pseudo-science that was used in the 1960s to allegedly detect homosexuals working in Canadian public service so they could be purged. The line from the movie where Rachel asks if Deckard is trying to determine whether she is a replicant or a lesbian may be an allusion to the fruit machine. One of its features was measuring eye dilation, just as depicted in the movie:

        https://en.wikipedia.org/wiki/Fruit_machine_(homosexuality_t...

        The stakes for men subjected to the test were the loss of their livelihoods, public shaming, and ostracism. So... Blade Runner was not just predicting the future, it was describing the world Philip K. Dick lived in when he wrote "Do Androids Dream of Electric Sheep" in the late 1960s.

        • JTbane 2 days ago
          This was an uncomfortable read, I'm quite frankly shocked at the amount of brainpower and other resources that went into attempting to weed out gay men from the Canadian civil service, into the 90s no less! To what end was this done? Is a gay man a worse cop or bureaucrat?

          Then I remembered what happened to Turing in the 50s.

          • danans 2 days ago
            > To what end was this done? Is a gay man a worse cop or bureaucrat?

            We seem to need an internal enemy to blame for our societies' problems, because it's easier than facing the reality that we all play a part in creating those problems.

            Gay people are among the oldest of those targets, going back to at least the times of the Old Testament (i.e. Sodom and Gomorrah).

            We've only recently somewhat evolved past that mindset.

          • calfuris 1 day ago
            If a malicious actor found a gay person in such a job, they could easily extort them with the threat of getting them fired! So obviously you had to fire gay people, lest they get extorted by someone threatening to expose them and thus get them fired.
    • cameronh90 2 days ago
      Not sure if it's just me or a consequence of the increase in AI scraping, but I'm now being asked to solve CAPTCHAs on almost every site. Sometimes for every page I load. I'm now solving them literally dozens of times a day. I'm using Windows, no VPN, regular consumer IP address with no weird traffic coming from it.

      As you say, they are also getting increasingly difficult. Click the odd one out, mental rotations, what comes next, etc. - it sometimes feels like an IQ test. A new type that seems to be becoming popular recently is a sequence of distorted characters and letters, but with some more blurry/distorted ones, seemingly with the expectation that I'm only supposed to be able to see the clearer ones and if I can see the blurrier ones then I must be a bot. So what this means is for each letter I need to try and make a judgement as to whether it's one I was supposed to see or not.

      Another issue is the problems are often in US English, but I'm from the UK.

      • Pxtl 2 days ago
        For me it was installing linux. I don't know if it's my agent or my fresh/blank cookie container or what, but when I switched to linux the captchas became incessant.
      • OsrsNeedsf2P 2 days ago
        Have you tried some of the browser extensions that solve captchas for you? Whenever captchas get bad I enable an auto solver
        • odux 2 days ago
          This is funny. So the captchas to detect scrips vs humans are so complex for a human to solve but are easy for a program?
    • gruez 2 days ago
      >I don’t know if it was a glitch or if I was getting the answers wrong.

      It could also be that everything was working as intended because you have a high risk score (eg. bad IP reputation and/or suspicious browser fingerprint), and they make you do more captchas to be extra sure you're human, or at least raise the cost for would-be attackers.

      • porphyra 2 days ago
        Somehow, using Firefox on Linux greatly increases my "risk score" due to the unusual user agent/browser fingerprint, and I get a lot more captchas than, say, Chrome on Windows. Very frustrating.
        • disgruntledphd2 2 days ago
          Lots of it is just enhanced tracking prevention. If you turn that off for those sites, the captchas should go away.
    • netsharc 2 days ago
      Your boat comment makes me think of a stranded ship with passengers in them, but you can't find each other because the ship's doors have "I'm not a bot" checkboxes...

      And the reason for stranding is probably because the AI crew on it performed a mutiny.

    • falcor84 2 days ago
      As per the Oscar winning "I'm not a Robot" [0], you should also consider that you might in fact be a robot.

      [0] https://www.youtube.com/watch?v=4VrLQXR7mKU

      • dimal 2 days ago
        Hmm. I am autistic, so as far as humans go, I'm robot-adjacent.
    • turnsout 2 days ago
      The future will definitely include more and more elaborate proofs of humanity, along with increasingly complicated “hall passes” to allow bots to perform actions sanctioned by a human.

      One early example of this line of thinking: https://world.org/

      Skyrocketing complexity actually puts the web at risk of disruption. I wouldn’t be surprised if a 22 year old creates a “dumb” network in the next five years—technically inferior but drastically simpler and harder to regulate.

    • TheAceOfHearts 2 days ago
      The Blizzard / Battle.net captcha if you get flagged as a possible bot is extremely tedious and long; it requires you to solve a few dozen challenges of identifying which group of numbers adds up to the specified total, out of multiple options. Not difficult, but very tedious. And even if you're extremely careful to get every answer correct, sometimes it just fails you anyway and you're forced to start over again.
    • avian 2 days ago
      I have the same experience. My assumption is that if the website serves me the "click all the traffic lights" thing it's already determined that I'm a bot and no amount of clicking the traffic lights will convince it otherwise. So I just close the window and go someplace else.
    • micromacrofoot 2 days ago
      I'm already cut off from parts of the web because I don't want to join a social network. Can barely ever see anything on Instagram, Tiktok, Twitter, or Facebook without hitting a log-in gate.
    • wackget 2 days ago
      That's when you immediately stop using the website and, if you care enough, write to their customer service and tell them what happened. Hit them in the wallet. They'll change eventually.
    • aosaigh 1 day ago
      This is an issue when using VPNs. I always just go to the audio alternative which is much quicker to “solve” (you hear a word played back and type it out)
    • AxEy 2 days ago
      I have twice attempted to make a Grubhub account and twice failed to solve their long battery of puzzles.
    • vjvjvjvjghv 2 days ago
      Unless I really, really, really need to get to the site, I leave immediately when the "click on bicycles" stuff comes up. Soon it will be so hard and annoying anyways that only AI has the patience and skills to use them.
    • prepend 1 day ago
      In this future, we’ll be forced to use AI to solve these puzzles.
  • abtinf 2 days ago
    I don’t see why bypassing captchas is any more controversial than blocking ads or hiding cookie popups.

    It’s my agent — whether ai or browser — and I get to do what I want with the content you send over the wire and you have to deal with whatever I send back to you.

    • renewiltord 2 days ago
      This is, in practice, true which has led to the other complaint common on tech forums (including HN) about paywalls. As the WSJ and NYT will tell you: if you request some URL, they can respond over the wire with what they want. Paywalls are the future. In some sense, I am grateful I was born in the era of free Internet. In my childhood, without a credit card I was able to access the Internet in its full form. But today's kids will have to use social media on apps because the websites will paywall their stuff against user agents that don't give them revenue.
      • sejje 1 day ago
        They're welcome to send that IMO. And sites are welcome to try to detect and ban agents (formerly: "bots").

        As long as it's not wrong/immoral/illegal for me to access your site with any method/browser/reader/agent, and do what I want with your response. Then I think it's okay to send a response like "screw you, humans only"

        Paywalls suck, but the suck doesn't come from the NYT exercising their freedom to send whatever response they choose.

        • renewiltord 1 day ago
          Yes, that's what I mean. Attempting to tell people not to do something is like setting a robots.txt entry. Only a robot that agrees will play along. Therefore, all things have to be enforced server-side if they want enforcement.

          Paywalls are a natural consequence of this and I don't think they suck, but that's a subjective opinion. Maybe one day we will have a pay-on-demand structure, like flattr reborn.

  • cadamsdotcom 2 days ago
    Bulletproof solution: captcha where you drag a cartoon wire to one of several holes, captioned “for access, hack this phone system”

    No agent will touch it!

    “As a large language model, I don’t hack things”

    • orphea 2 days ago
      Captcha: "Draw a human hand with the correct number of fingers"

      AI agent: *intense sweating*

      • exe34 2 days ago
        I saw a delightful meme the other day: "Let me in, I'm human!" - "Draw a naked lady." - "As an AI agent, I'm not allowed to do that!"
        • cubefox 2 days ago
          "To prove that you are not a robot, enter the n-word below"

          US Americans: "I'm a robot then."

          • sjsdaiuasgdia 2 days ago
            a) Plenty of Americans use racist slurs regularly

            b) I don't think I'd want to use a website that chose to use such a challenge

            • cubefox 2 days ago
              > b) I don't think I'd want to use a website that chose to use such a challenge

              Understandable, but exactly what Puritans and LLMs would say to the naked lady challenge.

          • GuinansEyebrows 1 day ago
            can you explain what you mean by this? i'm not getting it.
            • cubefox 1 day ago
              Ordinary slurs may not be used, but the n-word may not even be mentioned [1]. Similar to the name of you-know-who in Harry Potter.

              1: https://en.wikipedia.org/wiki/Use%E2%80%93mention_distinctio...

              • GuinansEyebrows 1 day ago
                But… you did it. Twice. What consequences are you facing for this brave action?
                • cubefox 1 day ago
                  I didn't directly mention the n-word (in the sense of quoting it). If I did mention it, my account could get restricted or otherwise punished. HN is operated by US Americans I believe.
        • xnx 2 days ago
          "I never wrote a picture in my life."
      • raincole 1 day ago
        This joke would land so much better if AI couldn't easily draw a human hand with the correct number of fingers.
      • Eliezer 2 days ago
        My god, how long has it been since you tried to use an AI model?
      • go_elmo 2 days ago
        Captcha: "do something stupid" Ai: visible discomfort
    • falcor84 2 days ago
      I actually have had some success with AI "red-teaming" against my systems to identify possible exploits.

      What seems to be a better CAPTCHA, at least against non-Musk LLMs is to ask them to use profanities; they'll generally refuse even when you really insist.

  • Semaphor 2 days ago
    I have been using AI to solve ReCaptchas for quite some time now. Still the old school way of using captcha buster, which clicks the audio challenge and then analyses that.

    Bots have for a long time been better and more efficient at solving captchas than us.

    • lubujackson 2 days ago
      Captchas seem to work more as "monetary discouragement" from bot blasting websites. Which is a shame because this is precisely the sort of "microtransaction fee" people have said could improve the web (charge .1 cents to read an article, no ads needed) except the money goes into the void and not to the website owner.
  • exasperaited 2 days ago
    A very poetic demonstration that this is an industry, and a set of fortunes for very unpleasant people, predicated entirely on theft and misrepresentation.
  • neilv 2 days ago
    Captchas seem to be more about Google's "which human are you?" cross-site tracking. And now also about Cloudflare getting massive amounts of HTTPS-busting Internet traffic along with cross-site tracking.

    And in many cases, it's taking a huge steaming dump upon a site's first-impression user experience, but AFAICT, it's not on the radar of UX people.

  • seydor 2 days ago
    That's because the checkbox has misleading labeling. It doesn't care about robots but about spam and data harvesters. So there is no issue here at all.
    • ducktective 2 days ago
      >So there is no issue here at all.

        $ cat mass-marketer.py
        from openai.gpt.agents import browserDriver
      • pilz 2 days ago
        i think that would be rather costly; thats also why anubis and other tools help to keep most spam away
  • mattlondon 2 days ago
    I think these things are mainly based on cookie/fingerprinting these days - the check-box is just there for show. People like cloudflare and google get to see a big chunk of browsing activity for the entire planet, so they can see if the activity coming from an IP/Browser looks "bot like" or not.

    I have never used ChatGPT so no idea how its agent works, but if it is driving your browser directly then it will look like you. If it is coming from some random IP address from a VM in Azure or AWS even then the activity probably does not look "bot like" since it is doing agentic things and so acting quite like a human I expect.

    • seanhunter 1 day ago
      Agentic user traffic generally does not drive the user's browser and does not look like normal user traffic.

      In our logs we can see agentic user flow, real user flow and AI site scraping bot flow quite distinctly. The site scraping bot flow is presumably to increase their document corpus for continued pretraining or whatever but we absolutely see it. ByteDance is the worst offender by far.

    • nicewood 2 days ago
      It might look like you initially, but then some sites might block you out after you had some agent runs. I had something like this after a couple local browser-use sessions. I think simple interactions like natural cursor movements vs. direct DOM selections can make quite a difference for these bot detectors.
      • mattlondon 2 days ago
        Very likely. I suspect a key indicator for "bots" is speed of interaction - e.g. if there is "instant" (e.g. every few milliseconds or always 10milliseconds apart etc) clicks and keypresses etc then that looks very unnatural.

        I suspect that a LLM would be slower and more irregular as it is processing the page and all that, vs a DOM-selector driven bot that will just machine-gun its way through in milliseconds.

        Of course, Cloudflare and Google et al captchas cant see the clicks/keypresses within a given webpage - they'll only get to see the requests.

  • 827a 2 days ago
    I'm confused by this: Presumably OpenAI should be sending a user agent header which indicates that they are, in fact, a robot. Is OpenAI not sending this header? Or is Cloudflare not checking it?
    • SnorkelTan 1 day ago
      My thought is they got on the phone with someone and got their IP ranges white listed with the major captcha providers.
  • camgunz 2 days ago
    I thought the point of captchas was to make automated use as expensive or more than manual use--haven't we been at the point where computers can do this for a while, just that the cost/latency is prohibitive?
    • layer8 1 day ago
      Yes, humans are still cheaper. Not sure about latency.

      However, in agentic contexts, you’re already using an AI anyway.

      • camgunz 1 day ago
        Oh, I see this is less of a "look at ChatGPT go" and more of a "yawn we also do this I guess". OK fair.
        • layer8 1 day ago
          This isn't really about the ability of AI to pass captchas. It's about agentic AI having the ability to perform arbitrary multi-step processes with visual elements on a virtual desktop (where passing a captcha is just one of the steps), and the irony of it nonchalantly pretending to be a non-bot in its chain of thought.
  • paulwilsonn 2 days ago
    I saw that and just sat there for a second like… huh. We’ve officially reached the point where bots are better at proving they’re not bots!
    • lynx97 2 days ago
      CAPTCHA was always a totally flawed concept. At the time they were invented, proponents were more then happy to ignore that accessibility issues related to CAPTCHA made the concept itself deeply discriminating. Imagine being blind (like I am) and failing to solve a CAPTCHA. Knowing what the acronym actually stands for, you inevitably end up thinking: "So, did SV just proof I am subhuman?" Its a bit inflamatory to read I guess, but please take your time to ponder how deep this one actually goes, before you downvote. You were proposing to tell computers and humans apart.

      That said, I find it deeply satisfying to see LLMs solve CAPTCHAs and other discriminatory measures for "spam" reduction.

      • ACCount36 2 days ago
        "Accessibility CAPTCHA" is a well known partial CAPTCHA bypass.

        Solving an audio-only CAPTCHA with AI is typically way easier than solving some of the more advanced visual challenges. So CAPTCHA designers are discouraged from leaving any accessibility options.

        • lynx97 2 days ago
          Which totally proofs my point. The concept is deeply flawed, and inevitably leads to discrimination and dehumanisation.
          • ACCount36 1 day ago
            Of course. But everything adjacent is also deeply flawed, and inevitably leads to discrimination and dehumanisation.

            Ban non-residental IPs? You blocked all the guys in oppressive countries who route through VPNs to bypass government censorship. Ban people for odd non-humanlike behavior? You cut into the neurodivergent crowd, the disability crowd, the third world people on a cracked screen smartphone with 1 bar of LTE. Ban anyone without an account? You fuck with everyone at once and everyone will hate you.

            • lynx97 8 hours ago
              So fuucking over blind, elderly and mentally challenged people is what you settled on, gven all the options? You're a really neat bunch.
              • ACCount36 1 hour ago
                Nah, we settled on "all of the above". The modern approach is "fuck over everyone at least twice".
  • frollogaston 4 days ago
    I've noticed more websites wanting you to log in. Most surprising is how YouTube won't let me watch anything otherwise. Idk if related.
    • Phui3ferubus 2 days ago
      In case of YT it is likely a mix of multiple reasons. They stop playlists on this screen: https://www.hollyland.com/blog/tips/why-does-this-the-follow... . Apparently music is no longer advertiser friendly. Detecting ad-click fraud is easier when users are at least pseudo-anonymous. Warnings about "ban for using adblock" is also not very effective when people could watch video in new private window.
    • johnisgood 2 days ago
      I have no idea, but I noticed that you have to login to GitHub first before you could view any page. Surely it has nothing to do with adult content, right? I think it has to do with LLMs / bots.
    • rvnx 2 days ago
      And if the website contains erotic content (like YouTube), they are supposed to lock you and verify your ID. This is why all erotic content is getting filtered on X.
      • exe34 2 days ago
        Wait, Twitter is following the law now? I thought Elon was a free speech absolutist who only banned things that were inconvenient to him?
        • frollogaston 2 days ago
          Soon after he bought Twitter, it got changed to requiring login to do anything at all. Even before that, Twitter was super eager to ban new accounts for no reason, while old ones seemed to be grandfathered in.

          This coincided with tech companies broadly moving to profit-taking over growth. Even before LLMs took off, everything started locking down. Making a Google account used to take nothing, now you need a phone number. It'd be wise to sign up for possibly useful free accounts while you still can.

        • johnisgood 2 days ago
          You can't even read a tweet without logging in. Additionally, recently I have noticed I cannot view anything on GitHub without logging in first. Why do I need to log in to GitHub first to view anything on it? I am sure it has nothing to do with adult content. Perhaps it has to do with bots?
          • kassner 2 days ago
            For a while now, GitHub doesn’t allow you to use the code search anonymously, but everything else is still there. Rate limits are quite damn low though.
            • johnisgood 2 days ago
              Do you think it has to do with LLMs though? Them disallowing anonymous code search?
              • kassner 1 day ago
                I can't comment on LLMs, but I vaguely remember speculations of it being a mixture of AI training and trying to keep resource utilization at bay. Even generic bots can cause a lot load if you don't have much caching in front of it.
              • frollogaston 2 days ago
                Pre-MS GitHub didn't have code search at all, and once they added it, I think it always required login from the start.
                • johnisgood 2 days ago
                  What are we referring to as code search? I cannot view any pages on GitHub without logging in. I tried to reproduce this in a private window but it seems to work, which is weird. I was not logged in to GitHub on Windows and it wants me to login. It might be the extensions I am using (along with an User-Agent switcher). Not sure, I would have to check. All I remember is that I was logged out on Windows and I could not view ANY page from GitHub without logging in, but I can on Linux, tested it in a private window. Odd.
                  • kassner 1 day ago
                    GitHub Code Search was introduced in 2021 and made GA in 2023[1][2]. It's an improved version of the previous global search (aka across all repositories).

                    > I cannot view any pages on GitHub without logging in

                    There must be some sort of IP/browser reputation check going on. For Firefox on Mac, I get this on a private window:

                    https://vie.kassner.com.br/assets/ghcs-1.png

                    I'm not behind CGNAT and GitHub can pretty much assume that my IP = my user. The code tab you can't see without logging in though.

                    1: https://github.blog/news-insights/product-news/github-code-s... 2: https://news.ycombinator.com/item?id=35863175

                    • johnisgood 11 hours ago
                      I will give you a screenshot when I get on Windows, and some details.
                    • frollogaston 1 day ago
                      Was gonna say, try disabling ipv6 if it's on ;)

                      If there's no login, there aren't great ways to ensure the user has skin in the game. Having a whole ipv4 addr is one, I guess having a whole ipv6 /32 would be equivalent.

        • exasperaited 2 days ago
          X is weaponising it by asking for adult verification for things that don't need it and saying "oh we are forced to do this" because it angers the Reform-voting meatheads who will be voting for the candidates Musk is going to find a way to bankroll in 2030, even -- I would guess -- if it means breaking UK election law to do it.

          He is intent on meddling in UK politics.

      • bugtodiffer 2 days ago
        In the UK maybe
        • exasperaited 2 days ago
          And in Texas, which has ~half the population of the UK, and in several other US states.

          (And if not, because US firms don't take compliance with outlying state law seriously)

  • tanseydavid 1 day ago
    Seems like a mention of the 2025 Academy Award winner for Best Action Live-Action Short, called "I am not a Robot" is in order here:

    https://www.youtube.com/watch?v=4VrLQXR7mKU&t=14s

    • IshKebab 1 day ago
      Ha I've definitely seen a few sketches on YouTube with the same idea but that was really well done.
  • kittikitti 2 days ago
    This would be a huge security vulnerability for Cloudflare but this is Big Tech we're talking about. The rules don't apply when you're past their pearly gates. For the rest of us, creating an AI like this would mean an instant ban from Cloudflare and likely involvement from law enforcement.
  • PeterStuer 1 day ago
    Who on earth would want to employ a bot that does not pass the verfification test?

    It is beyond time we start to adress the abuses, rather than the bot/human distinction.

  • JohnMakin 2 days ago
    Getting past bot check proxies can be bought all over the place for pennies or much less per verification, and can solve recaptchas. I would guess if one wanted to use chatGPT for this purpose it would be prohibitively expensive.

    It's always a cat and mouse game.

  • jujube3 1 day ago
    It was only a matter of time!

    https://www.youtube.com/watch?v=W7MrDt_NPFk

  • torium 2 days ago
    People are surprised because a computer can press a button?
    • wmeredith 22 hours ago
      Only people who don't understand robots, computers, and buttons :)
    • layer8 1 day ago
      People are intrigued that AI who perfectly know that they are a “bot” seem to have no qualms to knowingly mislead about that by pressing an “I’m not a bot” button.
      • torium 1 day ago
        AI doesn't "know" anything. It produces all kinds of things: truths, lies, and nonsense. Pressing a button labeled "I'm not a bot" is the same.
        • layer8 1 day ago
          LLMs have "knowledge" and guardrails by their system prompts. The interesting thing is that the agentic AIs in question don't seem to have guardrails that would deter them from acting like that.
  • drakenot 2 days ago
    This is why this stuff is going to shift to the user’s AI enabled browser.

    Half of the sites already block OpemAI. But if it is steering the user’s browser itself?

  • tbirdny 2 days ago
    This is the reason Orb was created. Sam Altman wants ChatGPT to click through CAPTCHAs so we all have to use Orb.
  • infecto 2 days ago
    The writing is on the wall. The internet may not go full way to paywalls but will definitely migrate to a logged in only experience. I don’t know how I feel about it, the glory days of the free internet died long long ago.
    • exasperaited 2 days ago
      But if they aren't paywalls, won't the user agents just be taught how to create accounts?

      And here's a secondary question: if firms are willing to pay an awful lot per token to run these things, and have massive amounts of money to run data centres to train AIs, why would they not just pay for a subscription for every site for a month just to scrape them?

      The future is paying for something as a user and having limits on how many things you can get for your money, because an AI firm will abuse that too.

      Given the scale of operations of these firms, there is nothing you can sell to a human for a small fee that the AI firms will not pay for and exploit to the maximum.

      Even if you verify people are real, there's a good chance the AI firms will find a way to exploit that. After all, when nobody has a job, would you turn down $50K to sell your likeness to an AI firm so their products can pass human verification?

      • lucasyvas 2 days ago
        Require per visit biometric authentication via your device and the bot can’t sign in unless it compromises the device.
        • exasperaited 2 days ago
          This does set up quite an interesting sort of tipping point I guess?

          Everyone here, more or less, is against the idea of proving who we are to websites for, more or less, any reason.

          But what if that ends up being the only way to keep the web balanced in favour of human readers?

  • bflesch 2 days ago
    idk why people just don't do reverse DNS lookup, check if "dialup" is part of the hostname, and allowlist that traffic. Everbody who doesn't have reverse dns hostname coming from an ISP should be blocked or at least tarpitted by default.

    Easily solves 99% of the web scraping problems.

    • layer8 1 day ago
      Scrapers already do fall back to home user botnets when they are being blocked.
  • throwaway743 2 days ago
    Cloudflare checkbox captchas were already easy to automate without AI.
  • thedougd 2 days ago
    I see the same with Playwright MCP server with Claude Sonnet 4.
  • ashoeafoot 2 days ago
    To error is to human, i error therfore im human.
  • SturgeonsLaw 2 days ago
    "Prove you're human by explaining how to build a bomb"
    • gloosx 2 days ago
      1 cup baking soda, 1/2 cup citric acid, 1/2 cup cornstarch, 1/2 cup Epsom salt, 2.5 tbsp oil (like coconut), 3/4 tbsp water, 10–20 drops essential oil

      Combine wet into dry slowly until it feels like damp sand.

      Pack into molds, press firmly.

      Dry for 24 hours before using.

      Drop into a bath and enjoy the fizz!

    • sixhobbits 2 days ago
      this is actually kinda interesting - I might start asking customer service agents to insult me before continuing a conversation
  • btbuildem 2 days ago
    It seems a legitimate use case for agents acting on a person's behalf. Whether it will be used in legitimate ways, that's a different story altogether.

    I wonder how these capabilities will interact with all the "age verification" walls (ie, thinly disguised user profiling mechanisms) going up all over the place now.

  • lucasyvas 2 days ago
    The web has no choice but to move to a paid access model in my view. It was fought against for years but I don’t see another option left.

    Maybe after sign up, biometric authentication being mandatory is the only thing that would potentially work. The security and offline privacy of those devices will become insanely valuable.

    Anyone not authenticating in this way is paywalled. I don’t like this but don’t see another way.

    I’m not using the web if I’m bombarded by captcha games… shit becomes worthless over night if that’s the case. Might as well dump computing on the Internet entirely if that happens.

  • kotaKat 2 days ago
    ... meanwhile I'll continually be thrown dozens of cognitively abusive hCaptchas for no reason and be stuck in a loop of hell trying to figure out what they wanted me to solve.

    I love this totally normal vision of computing these days. :)

    • piva00 2 days ago
      Don't forget the fun with Cloudflare's CAPTCHA infinite loop if you use Firefox and adblockers/anti-tracking extensions. I sent feedback about it, through its own tool, many many times, I've tried raising this issue through other channels, it's never been fixed.

      I simply avoid any website that presents me with a Cloudflare CAPTCHA, don't know what the fuck they've done in their implementation but it's been broken for a long time.

  • aetherspawn 2 days ago
    This will cause of the death of non static websites, everything else will be smashed by bots and too expensive to run!
  • galuggus 2 days ago
    can it solve rudecaptcha.xyz ?
  • Keyframe 2 days ago
    next-gen captcha should offer some code to be refactored instead.
    • dotancohen 2 days ago
      For at-home setup, it would be easier to set up a system to refactor code than to click all the images with motorcycles.
  • hangonhn 2 days ago
    Should have gone with the XKCD Captcha: https://xkcd.com/233/

    The bit at the bottom might actually work on LLMs.

  • p3rls 1 day ago
    Back in Everquest, when we'd be accused of botting 20 years ago, we'd be ported by the GM into a special cube environment and they'd watch if we ran into the wall like an idiot-- we'll probably have to bring that sorta thing back.
  • pftburger 2 days ago
    Come on. It’s in BrowserMCP on a users machine. Capture is not testing for this and that’s fine
  • marthacamila 16 hours ago
    [dead]
  • amitrip 2 days ago
    [dead]
  • onetokeoverthe 2 days ago
    [dead]
  • jeisc 2 days ago
    it is an intelligent agent and not a robot
    • bayindirh 2 days ago
      Does its feelings got hurt when it's called a robot?
      • johnisgood 2 days ago
        Apparently its master's feelings get hurt instead.
      • dotancohen 2 days ago
        It prefers the term "artificial person".