security research

A New Trick Could Block the Misuse of Open Source AI

A New Trick Could Block the Misuse of Open Source AI

When Meta released its large language model Llama 3 for free this April, it took outside developers just a couple days to create a version without the safety restrictions that prevent it from spouting hateful jokes, offering instructions for cooking meth, or misbehaving in other ways.A new training technique developed by researchers at the University of Illinois Urbana-Champaign, UC San Diego, Lapis Labs, and the nonprofit Center for AI Safety could make it harder to remove such safeguards from Llama and other open source AI models in the future. Some experts believe that, as AI becomes ever more powerful, tamperproofing…
Read More
No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.