Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Not necessarily. The insecurity from LLMs comes from the fact they’re a black box - what if it turns out that particular version can be easily tricked into giving out terrorism ideas. You could try to add safeguards on top, but they’ve already been bypassed if it has been used for something like that. You might just have to retrain it somehow to make it safe


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: