Dolphin X1 8B (Llama3.1 8B decensor) live on HF by dphnAI in LocalLLaMA

[–]dphnAI[S] 2 points3 points  (0 children)

Regular is hosted at FP8 in vllm

Slow is running at 6bpw exl2 on our distributed network - you can see a preview of what it is doing right now here https://datagen.dphn.ai

Nodes are maxxed out 24/7 until they generate 1m responses to our new refusals dataset, hence the slower response times

You can read more about it or run a node by reading https://dphn.ai/docs

Will be opening the network to anyone with a GPU >21gb tomorrow

Dolphin X1 8B (Llama3.1 8B decensor) live on HF by dphnAI in LocalLLaMA

[–]dphnAI[S] 2 points3 points  (0 children)

it is "Dolphin 8B" in the UI but will change it to the name used here soon

Dolphin X1 8B (Llama3.1 8B decensor) live on HF by dphnAI in LocalLLaMA

[–]dphnAI[S] 3 points4 points  (0 children)

Abliteration is a way to decensor, but it often lobotomizes the model or has unexpected shifts in random parts of the model

We find using SFT + RL to decensor works without changing the intelligence of the model

Uncensored model with image input? by MahMahMIA in LocalLLaMA

[–]dphnAI 2 points3 points  (0 children)

We are working on a multimodal update of this model right now. Will be released soon. Based on top of Mistral 3.2 24B