"a Large Language Model (#LLM) can be convinced to tell you how to build a bomb if you prime it with a few dozen less-harmful questions first"
https://techcrunch.com/2024/04/02/anthropic-researchers-wear-down-ai-ethics-with-repeated-questions/
@lupyuen It sounds like interrogation. Many of similar methods could possibly work too.
QOTO: Question Others to Teach Ourselves An inclusive, Academic Freedom, instance All cultures welcome. Hate speech and harassment strictly forbidden.