What’s actually going to kill LLMs is when the sweet VC money runs out and the vendors have to start charging what it actually costs to run.
This isn’t the case. Midjourney doesn’t receive any VC money since it has no investors and this ignores genned imagery made locally off your own rig.
You can run it on your own machine. It won’t work on a phone right now, but I guarantee chip manufacturers are working on a custom SOC right now which will be able to run a rudimentary local model.
Both apple and Google have integrated machine learning optimisations, specifically for running ML algorithms, into their processors.
As long as you have something optimized to run the model, it will work fairly well.
They don’t want to have independent ML chips, they want it baked into every processor.