Open Source AI trimmed for efficiency produced detailed bomb making instructions and other bad answers before retraining


  • UCR -researchers Retrain AI models to keep security intact when trimmed to smaller units
  • Changing the starting layer removes protection, retraining of recoveries blocked uncertain answers
  • Examination using LLAVA 1.5 showed reduced models that were denied dangerous prompt after training

Researchers at the University of California, Riverside, address the problem of weakened security in open source artificial intelligence models when adapted to smaller devices.

Since these systems are trimmed to run effectively on phones, cars or other low -power hardware, they may lose the protective measures designed to prevent them from producing offensive or dangerous material.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top