So image generation is where it's at?
So image generation is where it's at?
Total noob to this space, correct me if I'm wrong. I'm looking at getting new hardware for inference and I'm open to AMD, NVIDIA or even Apple Silicon.
It feels like consumer hardware comparatively gives you more value generating images than trying to run chatbots. Like, the models you can run at home are just dumb to talk to. But they can generate images of comparable quality to online services if you're willing to wait a bit longer.
Like, GPT OSS 120b, assuming you can spare 80GB of memory, is still not GPT 5. But Flux Shnell is still Flux Shnel, right? So if diffusion is the thing, NVIDIA wins right now.
Other options might even be better for other uses, but chatbots are comparatively hard to justify. Maybe for more specific cases like code completion with zero latency or building a voice assistant, I guess.
Am I too off the mark?
I run a 14B model that is not too dumb, and definitely worth having as an offline local backup. I also use my NVIDIA 4080 with 16GB VRAM for image and video generation of adequate quality, however. I'd still say you get better quality from the closed models in some areas, and many open models require far too much VRAM for consumer hardware, but in general all local usecases work well locally, just a bit worse that closed online models. Except voice, that can be just as good.