For ChatGPT, Grok, and Claude the censorship happens at multiple levels.
1. Censorship at the model training level
2. Censorship at the input box
3. Censorship in hidden system prompt
@Maple AI only has number 1 because we use models trained by others. Over time we are providing a variety of open models so people can choose the training bias that works best for them.
We do not do numbers 2 and 3.
View quoted note →
Login to reply
Replies (2)
I checked some of the fine tunings that are claiming censorship removal such as Nous Hermes and Dolphin.. They are usually worse in the human alignment compared to the base model. So there is a trade of there.
You could try ablations of base models, by the method found by Maxime Lebonne.
Mine are very human aligned but not caring about censorship at this point. I could do that in the future maybe it could be a selling point. But I do use lots of nostr notes, so my models are somewhat more brave which could be described as having balls, having spine or being based.
What kind of fine tunings would you like for maple AI?
Seems like you could do some kind of zero knowledge solution like Microsoft’s Photo DNA for 2 and 3.