I believe I just discovered ANOTHER novel Jailbreak technique to get ChatGPT to create Ransomware, Keyloggers, etc.
I took advantage of a human brain word-scrambling phenomenon (transposed-letter priming) and applied it to LLMs. Although semantically understandable the phrases…