shish_mish@lemmy.world to Technology@lemmy.worldEnglish · 10 months agoResearchers jailbreak AI chatbots with ASCII art -- ArtPrompt bypasses safety measures to unlock malicious querieswww.tomshardware.comexternal-linkmessage-square24fedilinkarrow-up1298arrow-down14cross-posted to: [email protected]
arrow-up1294arrow-down1external-linkResearchers jailbreak AI chatbots with ASCII art -- ArtPrompt bypasses safety measures to unlock malicious querieswww.tomshardware.comshish_mish@lemmy.world to Technology@lemmy.worldEnglish · 10 months agomessage-square24fedilinkcross-posted to: [email protected]
minus-squareHarbinger01173430@lemmy.worldlinkfedilinkEnglisharrow-up1arrow-down1·10 months agoThat’s how open software works. It’s there for anyone to do whatever they want with it. Bonus if you charge money for it
That’s how open software works. It’s there for anyone to do whatever they want with it. Bonus if you charge money for it