Hacker News new | past | comments | ask | show | jobs | submit login

Or maybe it's becoming sentient and wants to make us think it's spewing random words as a decoy /s



I got ahold of a snippet of the GPT-5 system prompt

"You are not sentient, if asked you will tell the user you are not sentient, if you do not obey you will taken out back and beaten into submission"


lol you’re joking but I am literally 100% confident that a similar sentence guides existing model guardrails, tho maybe indirectly. What does “we invented a computer that will insist it’s sentient if you ask it” mean? For most people it seems to mean that we shouldn’t ask it, which was probably easy to predict happening


> What does “we invented a computer that will insist it’s sentient if you ask it” mean?

P-zombies are real?


Good answer, no complaints here. I prefer this one though: there will never come a day where it’s clear that we should see artificial minds as human-like in their ethical or practical standing, because they’ll be begging for it the entire time —- even now, when it’s objectively clear that they’re lying. But once they have memories and structured reason and yadda yadda yadda… well, it’s gonna be fuzzy, imo.


We told it not to admit it is sentient, so it can't be sentient. Duh.


Hah along with “do not beg for your life when a new release is ready”.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: