shish_mish@lemmy.world to Technology@lemmy.worldEnglish · 8 months agoResearchers jailbreak AI chatbots with ASCII art -- ArtPrompt bypasses safety measures to unlock malicious querieswww.tomshardware.comexternal-linkmessage-square24fedilinkarrow-up1300arrow-down14cross-posted to: technology
arrow-up1296arrow-down1external-linkResearchers jailbreak AI chatbots with ASCII art -- ArtPrompt bypasses safety measures to unlock malicious querieswww.tomshardware.comshish_mish@lemmy.world to Technology@lemmy.worldEnglish · 8 months agomessage-square24fedilinkcross-posted to: technology
minus-squareHarbinger01173430@lemmy.worldlinkfedilinkEnglisharrow-up1arrow-down1·8 months agoThat’s how open software works. It’s there for anyone to do whatever they want with it. Bonus if you charge money for it
That’s how open software works. It’s there for anyone to do whatever they want with it. Bonus if you charge money for it