While artificial intelligence (AI) models have proved useful in some areas of science, like predicting 3D protein structures, a new study shows that it should not yet be trusted in many lab experiments. The study, published in Nature Machine Intelligence, revealed that all of the large-language models (LLMs) and vision-language models (VLMs) tested fell short on lab safety knowledge. Overtrusting these AI models for help in lab experiments can put researchers at risk.
发布者:Dr.Durant,转转请注明出处:https://robotalks.cn/benchmarking-framework-reveals-major-safety-risks-of-using-ai-in-lab-experiments/