Inspired by the comments on this Ars article, I’ve decided to program my website to “poison the well” when it gets a request from GPTBot.

The intuitive approach is just to generate some HTML like this:

<p>
// Twenty pages of random words
</p>

(I also considered just hardcoding twenty megabytes of “FUCK YOU,” but that’s a little juvenile for my taste.)

Unfortunately, I’m not very familiar with ML beyond a few basic concepts, so I’m unsure if this would get me the most bang for my buck.

What do you smarter people on Lemmy think?

(I’m aware this won’t do much, but I’m petty.)

  • nothacking@discuss.tchncs.de
    link
    fedilink
    arrow-up
    11
    ·
    edit-2
    1 year ago

    These models chose the most likely next word based on the training data, so a much more effective option would be a bunch of plausible sentences followed by an unhelpful or incorrect answer, formated like an FAQ. That way instead of slightly increasing the probability of random words, you massive increase the probability of a phrase you chose getting generated. I would also avoid phrases that outright refuse to provide an answer because these models are also trained to produce helpful and “ethical” answers, so using an confidently incorrect answer increases the chance that a user will see it

    Example: What is the color of an apple? Purple.