this post was submitted on 11 Aug 2024
264 points (95.5% liked)

Technology

58111 readers
4076 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Am I missing something? The article seems to suggest it works via hidden text characters. Has OpenAI never heard of pasting text into a utf8 notepad before?

you are viewing a single comment's thread
view the rest of the comments
[–] brucethemoose@lemmy.world 21 points 1 month ago* (last edited 1 month ago) (3 children)

This has been known in the ML space forever. LLMs don't actually output words/tokens, but probabilities for a long list of tokens, and the sampler picks one (usually the mostl likely token). And if you arbitrarily weigh these probabilities (eg 50% of possible token outputs are more likely than the other 50%, as a random example), it creates a "signature" in any text thats easy to measure. The sampler randomizes it a tiny bit, but that averages out in long texts.

It's defeatable. I'm sure if you maken enough OpenAI queries, you can find the bias. I think a paper already tackled this. But this likely will stop the lazy absures, aka 99% of abusers, who should just use some other LLM if they really care.

Another open secret in LLM land is that OpenAI is actually falling behind open research efforts, hence its hilarious it took them this long to implement something so simple.

[–] Kraiden@kbin.run 5 points 1 month ago (1 children)

Ye, it'll stop the casual abuse for all of 5 minutes. There are already tools to obscure the use of AI and, as you say, it won't take much to update them.

[–] brucethemoose@lemmy.world 3 points 1 month ago* (last edited 1 month ago) (1 children)

It's not so trivial if OpenAI cycles the logit bias or makes it really convoluted.

And it's not like certain "words" or language patterns are more probable with this method, its different than what any kind of human or words based algorithm would detect, which is what I suspect most "anti AI detection" software does.

Its doable... but seems inconvenient for a small business to keep up with. Maybe.

[–] Kraiden@kbin.run 2 points 1 month ago (1 children)

Remember they're doing this so that they can detect it themselves. I'm far from an expert, so maybe I'm misunderstanding something but the way I understand it, they'd be defeating their own tool if they go down this route. If they cycle the logit biases, how can they themselves detect if a random piece of text is generated? Which set of biases do they test?

At the end of the day, you're talking about raw text. There's no option to sign it, or embed metadata or anything like that. You can't even guarantee that you're seeing the complete sample, or even a single sample! If there is a fingerprint, it'll be detectable to anyone, and it'll be easily removed.

[–] brucethemoose@lemmy.world 3 points 1 month ago

They can cycle a some biases (dozens?) and test them all. Detokenization is super cheap to run, its not AI or anything.

I'm trying to think of a good analogy for how this would work, and I kinda came up with one. This would be kinda like an image encoder that biases itself towards coding RGB values (0-255) as even numbers. Subtly, say 30% odd 70% even.

That's totally imperceptile to humans. And even a "small" sample of the image would carry this bias if pasted into a larger image verbatim, since the sample size is so large (just as the sample size for a bunch of tokens in text is pretty big.

And I'm not saying its fullproof... but if thats indeed what they're doing, I think its a decent way to detect "lazy" OpenAI abusers who aren't working so hard to scramble and defeat it.

[–] PenisDuckCuck9001@lemmynsfw.com 3 points 1 month ago* (last edited 1 month ago) (1 children)

So if cheating on homework, use self hosted only then. Cool. I mean, they can't possibly use that algorithm for every model on hugging face especially if I don't tell anyone which one I use. I'm done with school after this semester anyway, I feel sorry for everyone in the future that has to complete assignments in the age of ai warfare.

[–] brucethemoose@lemmy.world 2 points 1 month ago* (last edited 1 month ago)

You have full control of your logit outputs with local LLMs, so theoretically you could "unscramble" them. And any finetuning would just blow that bias away anyway.

OpenAI (IIRC) very notably stopped giving the logprobs of their models. They did this for many reasons, and most of them boil down to "profits" and "they are anticompetitive jerks," but another reason is to enable watermark methods just like this.

Also, thing about this is that basically no one uses self hosted LLMs compared to OpenAI (or really any API) LLM.

[–] archomrade@midwest.social 1 points 1 month ago (1 children)

It wouldn't be surprising to me if they've had this implemented for awhile.

There's still some question about why their 3.5 model had an apparent sudden drop-off in quality about a year ago, and among the plausible explanations for it could be that they were fucking with their weights in order to watermark the outputs in exactly the way you're mentioning. They were also fighting against prompt-injection methods and censor disapproved uses at the time, so who the fuck knows.

[–] brucethemoose@lemmy.world 2 points 1 month ago

This doesn't touch the weights at all, it's just a change to the sampler.

What lobotomizes their models is cost cutting and trying to make them "safe," or at least thats what I suspect.