Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I always assumed that they were snake oil because the training objective is to get a model that writes like a human. AI detectors by definition are showing what does not sound like a human, so presumably people will train the models against the detectors until they no longer provide any signal.


The thing is, the LLM has a flaw: it is still fundamentally biased towards frequency.

AI detectors generally can take advantage of this and look for abnormal patterns in frequencies of specific words, phrases, or even specific grammatical constructs because the LLM -- by default -- is biased that way.

I'm not saying this is easy and certainly, LLMs can be tuned in many ways via instructions, context, and fine-tuning to mask this.


Couldn't the LLM though just randomly replace/reword things to cover up its frequency in "post"?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: