You generally want to use a model which has been fine tuned to work around the inbuilt censorship. There are plenty available on huggingface currently. It’s not a perfect solution, but works well enough for what it is.
I would suggest using the llama.cpp backend with a frontend of your choosing.