That's human-like intelligence at its finest. I am not being sarcastic, hear me out. If you told a person to give you 10 numbers at random, they can't. Everyone thinks randomness is easy, but it isn't ( see: random.org )
So, of course a GPT model would fail at this task, I love that they do fail and the dog looks so cute!!
I used to use Google assistant to spell words I couldn't remember the spelling of in my English classes (without looking at my phone) so the students could also hear the spelling out loud in a voice other than mine.
Me: "Hey Google, how do you spell millennium?"
GA: "Millennium is spelled M-I-L-L-E-N-N-I-U-M."
Now, I ask Gemini: "Hey Google, how do you spell millennium."
Gemini: "Millennium".
Most AI models out there are pretty brain dead as far as understanding goes, these types of things show the problems because it's abundantly clear it's getting it wrong. Makes you wonder how much it's getting wrong even when it isn't obvious.
For stuff like this to work correctly it must not be filtered through an MoE, it needs to be a direct prompt to a GenAI model that supports negative prompts.
Edit: I suppose a properly configured MoE with reasoning capabilities could probably do it