Download and install llamacpp from its github repository, go on huggingface.co and download one of the wizard vicuna uncensored GGUF models. It’s the most obedient and loyal one and will never refuse even the most ridiculous request. Use --threads option to specify more threads for higher speed. You’re welcome.