Discussion about this post

User's avatar
Limited Edition Jonathan's avatar

Omg- I'm glad I'm not the only one to see this. I mean, of course I tried to break it on day one (who doesn't?) but it was impossible, for me at least. I couldn't get it to tell me how to rig an election or build a... uh... sparkler.

I didn't play with it for more than a few hours since it couldn't do any serious work for me either. 🤷‍♂️

Expand full comment
Alistair Windsor's avatar

Sam Altman said “We have worked hard to mitigate the most serious safety issues, especially around biosecurity. gpt-oss models perform comparably to our frontier models on internal safety benchmarks.”

I think that OpenAI is very afraid that their open source model could be used for nefarious purposes.

Having said that, aligning any model is fraught with difficulties and every frontier model has quickly been jail broken. Closed source frontier models have the advantage that their queries can be preprocessed and their responses post-processed to look for violations of their safety policies in a way that open source models never can. Also, additional fine-tuning of open source models has been shown to relatively quickly bypass alignment.

If gpt-oss doesn’t produce objectionable answers and ends up not being widely used then I suspect Sam Altman will consider it a win. There is no real upside to OpenAI of wide-spread adoption of gpt-oss, but obvious downsides of having their model generate content that someone with wide distribution finds objectionable. Releasing the model has at least quieted those critics who pointed out that OpenAI never produced open source models.

Expand full comment

No posts