New study from EPFL shows that also one of the most current huge language versions (LLMs), regardless of going through security training, continue to be at risk to easy input adjustments that can trigger them to act in unintentional or hazardous means.
发布者:Dr.Durant,转转请注明出处:https://robotalks.cn/can-we-convince-ai-to-answer-harmful-requests/