Saturday July 27, 2024


Anthropic researchers wear down AI ethics with repeated questions
  Posted by: TechCrunch on Apr 2nd, 2024 8:33 PM

How do you get an AI to answer a question it’s not supposed to? There are many such “jailbreak” techniques, and Anthropic researchers just found a new one, in which a large language model can be convinced to tell you how to build a bomb if you prime it with a few dozen less-harmful questions […]

© 2024 TechCrunch. All rights reserved. For personal use only.



See Original Article At TechCrunch



View More Headlines