Skip to main content


A small number of samples can poison LLMs of any size:
anthropic.com/research/small-s…

"In a joint study with the UK AI Security Institute and the Alan Turing Institute, we found that as few as 250 malicious documents can produce a "#backdoor" vulnerability in a large language model—regardless of model size or training data volume."

Size does not matter: the #LLM edition. 😜

#AI #Claude #backdoors #malware #Anthropic

in reply to Karl Voit

So... can that be used as a weapon? My immediate thought is that if enough people put some of these documents on their public servers, maybe we can fight back those crawler bots that ignore robots.txt