If you run your own AI and watch how long it takes, how much it runs up the resources for a few seconds, then you might get an idea of what it's like hosting at least three copies of a multi-terabyte LLM, in memory, with much shorter response and a much bigger knowledge base (Gemini by Google), taking millions of prompts per minute. Then think of every company that's hosting major public AI services.
Then remember that the only things good that come out of AI are natural language inference for voice commands and slightly improved developer processes.
Hosting it is just a reminder of the rapid environmental, ecological and cultural destruction that is the AI bubble.
In summary: Perhaps, if the hoster wants it for streamlining their dev process. Otherwise it can be replaced with a far more efficient standard algorithmic program, which is what we had before.




That's cool, but wtf is it a conversation on Lemmy for