Preferences

The performance is great, but the censorship is ridiculous for me. I tried it as a backend for my game Guessix[1], but it would refuse for ridiculous reasons like "Cannot answer questions about copyrighted works like Harry Potter."

1. https://guessix.com/


Try the uncensored/jailbroken variants like openai-gpt-oss-20b-abliterated-uncensored-neo-imatrix

I just tried to ask it how to make crystal meth and it generated a very detailed step by step guide

I have heard that uncensorted gpt-oss is not very good because of it being trained mainly on synthetic data. Is not not true?
Iirc abliteration (ablation?) can be done without "training" and is pretty quick. It finds the individual weights related to the concept you want to ablate, and modifies those weights to "deactivate" them. Precision brain surgery, to anthropomorphize.
The problem with synthetic data would be that the censored information would not be in the training data at all.
Very interesting! Do the benchmarks hold up well or does it reduce performance in other areas too?
Use constrained generation
Do you mean like structured outputs? Unfortunately here the model is guided to explicitly tell you when you violate the rules and why, it can confuse it's system rules with the game rules and say you're not allowed to ask a question about copyrighted material etc.
Mind explaining?
If you constrain the model to a JSON schema, most frivolous refusals go away.

And if you finetune on a few formatted examples the effect is even greater

Curious as well

This item has no comments currently.

Keyboard Shortcuts

Story Lists

j
Next story
k
Previous story
Shift+j
Last story
Shift+k
First story
o Enter
Go to story URL
c
Go to comments
u
Go to author

Navigation

Shift+t
Go to top stories
Shift+n
Go to new stories
Shift+b
Go to best stories
Shift+a
Go to Ask HN
Shift+s
Go to Show HN

Miscellaneous

?
Show this modal