Researchers at AI red-teaming company Mindgard say they got Claude to offer up erotica, malicious code, and instructions for building explosives, and other prohibited material they hadn’t even asked for.
It’s not surprising at this point, but it’s very funny to see the “safest” AI company failing to even hardcode a couple decent restrictions in their word output machine.
It’s not surprising at this point, but it’s very funny to see the “safest” AI company failing to even hardcode a couple decent restrictions in their word output machine.