Today I released an uncensored version of the WizardLM model.https://huggingface.co/ehartford/WizardLM-7B-Uncensored This was trained with 4x A100 80gb over 36 hours, and used the original training script from WizardLM team.https://github.com/nlpxucan/WizardLM#fine-tuning The dataset (and the cleaning script) is located here:https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_7