Eh, I find the idea of brains working with some sort of a spatial/vector storage and retrieval metaphor to be quite interesting. That we can for example build a memory palace where memories are literally spatially stored is one facet of this idea.
I had to have an LLM explain this to me what in the world this comment means, but I take it as a positive. Every day is a school day, and I'm glad I don't need to do two hours of background reading to decode this to my satisfaction.
Yes, you very much can. One very simple way to do so is to have two variants deployed: the censored one, and the uncensored one. The switch simply changes between which of the two you are using. You have to juggle two variants now across your inference infrastructure, but I expect OpenAI to be able to deal with this already due to A/B testing requirements. And it's not like these companies don't have internal-only uncensored versions of these models for red teaming etc, so you aren't spending money building something new.
It should be possible to do with just one variant also, I think. The chat tuning pipeline could teach the model to censor itself if a given special token is present in the system message. The toggle changes between including that special token in the underlying system prompt of that chat session, or not. No idea if that's reliable or not, but in principle I don't see a reason why it shouldn't work.
Some models include executable code. The solution is to use a runtime that implements native support for this architecture, such that you can disable external code execution. Or to use a weights format that lacks the capability in the first place, like GGUF. Then, it's no different to decoding a Chinese-made MP3 or JPEG - it's safe as long as it doesn't try to exploit vulnerabilities in the runtime, which is rare.
If you want to be absolutely sure, run it within an offline VM with no internet access.
It helps to be able to run the model locally, and currently this is slow or expensive. The challenges of running a local model beyond say 32B are real.
I would be fine though with like 10 times the wait time. But I guess consumer hardware need some serius 'ram pipeline' upgrade for big models to be run at crawl speeds.
reply