As long as this allows running local, free software models I don't see the drawback of including this.
My main issue with ChatGPT and similar products is that they use my data to train their models. Running a model locally (like Llama) solves this problem, but running LLM models require extremely powerful GPUs, specially the bigger ones like Llama 70b.
So dedicated hardware for this is a nice thing for those that want it.
It requires powerful gpus yes but not always.
It depends a lot on how fast you want it to run. Microsoft and openai need powerful ai gpus because they have a lot of requests, data and want it to go fast. The dataset may also require to be stored in memory or gpu memory for fast access and use by the ai.
For Llama, it has been released as open source. And what is amazing about open source, is the community.
A Llama entirely in c++ has been created https://github.com/ggerganov/llama.cpp .
sure, a company that has used petabytes of data they do not own any rights of to train their models are totally excluding their own customers data when turning a switch off.
yeah, I totally trust OpenAI and Microsoft with my data. It's not like Microsoft is spying on me after turning of Windows telemetry either.