Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Nice read

> I'd be grateful for any pointers to an example where system developers (or someone else in a position to know) have verified the success of a prompt extraction.

You can try this yourself with any open source llm setup that lets you provide a system prompt no? Just give it a prompt, ask the model the prompt ,and see if it matches.

gpt-oss is trained to refuse so it wont share (you can provide system prompt on lmstudio)



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: