• BrianTheeBiscuiteer@lemmy.world
    link
    fedilink
    English
    arrow-up
    55
    ·
    9 months ago

    If it doesn’t get queried that’s the fault of the webscraper. You don’t need JS built into the robots.txt file either. Just add some line like:

    here-there-be-dragons.html
    

    Any client that hits that page (and maybe doesn’t pass a captcha check) gets banned. Or even better, they get a long stream of nonsense.

        • Aniki 🌱🌿
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          9 months ago

          That was really interesting. I always used urandom by practice and wondered what the difference was.

      • Aniki 🌱🌿
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        edit-2
        9 months ago

        I wonder if Nginx would just load random into memory until the kernel OOM kills it.

    • gravitas_deficiency@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      11
      arrow-down
      3
      ·
      9 months ago

      I actually love the data-poisoning approach. I think that sort of strategy is going to be an unfortunately necessary part of the future of the web.