Nah, I tried 20B and a bit of 120B. For the size, they suck, mostly because there’s a high chance they will randomly refuse anything you ask them unless it STEM or Code.
…And there are better models if all you need is STEM and Code.
Look around localllama and AI communities, they’re kinda a laughing stock even more than Llama 4.
Did you miss that open ai released the oss model few days prior to gpt5?
Larger model: https://huggingface.co/openai/gpt-oss-120b Smaller model: https://huggingface.co/openai/gpt-oss-20b
They seem to be quite good
Not saying that openai would be the good guys here, but I believe they are realizing that they are behind on local models.
I haven’t played with it too much yet but Qwen 3 seems better than GPt-OSS
Nah, I tried 20B and a bit of 120B. For the size, they suck, mostly because there’s a high chance they will randomly refuse anything you ask them unless it STEM or Code.
…And there are better models if all you need is STEM and Code.
Look around localllama and AI communities, they’re kinda a laughing stock even more than Llama 4.
In my limitrd testing it seemed relatively trigger-happy on refusals and the results were not impressive either. Maybe on par with 3.5?
Although it is fast at least.
https://mk.absturztau.be/notes/ab3gv6iygjam02uj