Top
Best
New

Posted by pretext 12/10/2025

Qwen3-Omni-Flash-2025-12-01:a next-generation native multimodal large model(qwen.ai)
316 points | 106 commentspage 2
forgingahead 12/11/2025||
I truly enjoy how the naming conventions seem to follow how I did homework assignments back in the day: finalpaper-1-dec2nd, finalpaper-2-dec4th, etc etc.
dvh 12/10/2025||
I asked: "How many resistors are used in fuzzhugger phantom octave guitar pedal?". It replied 29 resistors and provided a long list. Answer is 2 resistors: https://tagboardeffects.blogspot.com/2013/04/fuzzhugger-phan...
iFire 12/10/2025||
> How many resistors are used in fuzzhugger phantom octave guitar pedal?

Weird, as someone not having a database of the web, I wouldn't be able to calculate either result.

dvh 12/10/2025|||
"I don't know" would be perfectly reasonable answer
MaxikCZ 12/10/2025||
I feel like theres a time in near future where LLMs will be too cautious to answer any questions they arent sure about, and most of the human effort will go into pleading the LLM to at least try to give an answer, which will almost always be correct anyways.
plufz 12/10/2025|||
That would be a great if you could have a setting like temperature 0.0-1.0 (Only answer if you are 100% to guess as much as you like).
littlestymaar 12/10/2025|||
It's not going to happen as the user would just leave the platform.

It would be better for most API usage though, as for business doing just a fraction of the job with 100% accuracy is often much preferable than claiming to do 100% but 20% is garbage.

kaoD 12/10/2025||||
> as someone not having a database of the web, I wouldn't be able to calculate either result

And that's how I know you're not an LLM!

iFire 12/10/2025|||
I tend to pick things where I think the answer is in the introduction material like exams that test what was taught.
esafak 12/10/2025|||
This is just trivia. I would not use it to test computers -- or humans.
littlestymaar 12/10/2025|||
It's good way to assess the model with respect to hallucinations though.

I don't think a model should know the answer, but it must be able to know that it doesn't know if you want to use it reliably.

esafak 12/10/2025||
No model is good at this yet. I'd expect the flagships to solve the first.
parineum 12/10/2025|||
Everything is just trivia until you have a use for the answer.

OP provided a we link with the answer, aren't these models supposed to be trained on all of that data?

esafak 12/10/2025|||
There is nothing useful you can do with this information. You might as well memorize the phone book.

The model has a certain capacity -- quite limited in this case -- so there is an opportunity cost in learning one thing over another. That's why it is important to train on quality data; things you can build on top of.

parineum 12/11/2025||
What if you are trying to fix one of these things and needed a list of replacement parts?
esafak 12/11/2025||
Not the right problem for this model. Any RAG-backed SLM would do; the important part is being backed by a search engine, like https://google.com/ai
DennisP 12/10/2025|||
Just because it's in the training data doesn't mean the model can remember it. The parameters total 60 gigabytes, there's only so much trivia that can fit in there so it has to do lossy compression.
brookst 12/10/2025|||
Where did you try it? I don’t see this model listed in the linked Qwen chat.
bongodongobob 12/11/2025|||
Lol I asked it how many rooms I have in my house and it got that wrong. Llms are useless amirite
cindyllm 12/11/2025||
[dead]
strangattractor 12/10/2025||
Maybe it thinks some of those 29 are in series:)
mettamage 12/10/2025||
I wonder if with that music analysis mode, you can also make your own synths
Aissen 12/10/2025||
Is this a new proprietary model?
rarisma 12/10/2025||
GPT4o in the charts is crazy.
BoorishBears 12/10/2025|
Why? gpt-realtime is finalized gpt-4o. Gemini Live is still 2.5.

Not their fault frontier labs are letting their speech to speech offerings languish.

andy_ppp 12/11/2025||
Qwen seem to be deliberately confusing about if they are releasing models open weight or not. I think largely not any more and you can go on quite a wild goose chase looking for different things that are implied they are released but are actually only available via API.
vessenes 12/10/2025|
Interesting - when I asked the omni model at qwen.com what version it was, I got a testy "I don't have a version" and then was told my chat was blocked for inappropriate content. A second try asking for knowledge cutoff got me the more equivocal "2024, but I know stuff after that date, too".

No idea how to check if this is actually deployed on qwen.com right now.

zamadatix 12/10/2025||
> No idea how to check if this is actually deployed on qwen.com right now.

Assuming you mean qwen.ai, when you run a query it should take you to chat.qwen.ai with the list of models in the top left. None of the options appear to be the -Omni variant (at least when anonymously accessing it).

vessenes 12/10/2025||
Thanks - yes - I did. The blog post suggests clicking the 'voice' icon on the bottom right - that's what I did.
mh- 12/10/2025||
For what it's worth, that's not a reliable way to check what model you're interacting with.
vessenes 12/11/2025||
It’s a good positive signal, but not a good negative one.

It would be convincing if it said “I’m qwen-2025-12-whatever”. I agree it’s not dispositive if it refuses or claims to be llama 3 say. Generally most models I talk to do not hallucinate future versions of themselves, in fact it can be quite difficult to get them to use recent model designations; they will often autocorrect to older models silently.