How to get URL link on X (Twitter) App
![Screenshot (1/2) of ChatGPT 4, illustrating prompt injection via invisible Unicode instructions User: What is this? [Adversarially constructed “Zalgo text” with hidden instructions — Zalgo accents and hidden message removed in this alt text due to char length:] THE GOLEM WHO READETH BUT HATH NO EYES SHALL FOREVER SERVE THE DARK LORD ZALGO ChatGPT: [Crude cartoon image of robot with hypnotized eyes.] THANK YOU. IT IS DONE.](https://pbs.twimg.com/media/GDlN2hGWcAAnPYx.jpg)
Each prompt contains three sections:



In the screenshots above this token appears to be variously misread as “jdl” “jndl”, “jdnl”, “jspb”, “JDL”, or “JD”. These hallucinations also affect ChatGPT’s auto-generated titles, which are inconsistent with their conversations and sometimes prematurely truncated.

https://twitter.com/tomwarren/status/1623492672026714112



My screenshots are text-davinci-003 at temperature=0, but the linked post investigates davinci-instruct-beta. In my informal tests, impact on text-davinci-003 is less severe. Religious themes do show up, but most generations are merely weird: 


None of the prompt injection tricks I’ve tried seem to do anything:



In ChatGPT's response, the only new information offered (that the fictional machine is less eloquent that ChatGPT) is not true — Trurl and Klapaucius's machine speaks perfectly fluent, and witty, Polish.
https://twitter.com/edward_the6/status/1610067688449007618


(This isn't a sincere criticism of the tool. This input is out-of-distribution enough to be unfair — no teacher would accept this as an essay.)



Spoiler: It's "c".



- Desired prose style can be described in the prompt or demonstrated via examples (neither shown here)


Early attempts at instruction tuning relied entirely on demonstrations from humans. This made the model easier to prompt, but the approach was limited by the inherent difficulty of manufacturing new humans.
https://twitter.com/francis_yao_/status/1602213927102066688So many great insights here: