I have to think that most people won’t want to do local training.
It’s like Gentoo Linux. Yeah, you can compile everything with the exact optimal set of options for your kit, but at huge inefficiency when most use cases might be mostly served by two or three pre built options.
If you’re just running pre-made models, plenty of them will run on a 6900XT or whatever.
The Internet boom didn’t have the weird you’re-holding-it-wrong vibe too. Legit “It doesn’t help with my use case concerns” seem to all too often get answered with choruses of “but have you tried this week’s model? Have you spent enough time trying to play with it and tweak it to get something more like you want?” Don’t admit limits to the tech, just keep hitting the gacha.
I’ve had people say I’m not approaching AI in “good faith”. I say that you didn’t need “good faith” to see that Lotus 1-2-3 was more flexible and faster than tallying up inventory on paper, or that AltaVista was faster than browsing a card catalog.