IMHO Not a chance teams are running any LLM for anything critical:
* Training one is an insane cost, comparable to RBR catering costs per model trained, $4mil, 2048 A100 GPUs and 1 month to train LLaMa which only has 65B parameters
* Running one is equally nuts on costs, especially at scale.
* When it comes to really advanced topics, like math or physics, they are usually very wrong and you can eventually fix it with enough iterations on the answers. (
https://www.reddit.com/r/ChatGPT/commen ... d_in_math/ ,
https://cs.stanford.edu/~knuth/chatGPT20.txt )
* Lastly if they use the open ai product there is no expectation of privacy of the inputs, something that is a big no no for any team (Apple for example has forbidden its workforce from using it)
They could use something simple like LLaMa but that's not really very good.
The technology of this is evolving so rapidly that every few weeks at the moment there's a new discovery: bit quantization on LLaMa model to shrink it, or a new way to architect the model 4 days ago (
https://www.artisana.ai/articles/meta-a ... chitecture), without necessarily going in the security issues like prompt injection that are still being explored and have no real good practical solution yet.
LLM are language models, specialized tools from the teams will outperform LLMs except for open ended search tasks. For example they could use one of those models to test ideas provided they have a model trained on aero physics but it's entirely possible it spits out gibberish because it's a language model, not a math model.
Even with all this the tech behind LLM is insane and amazing.
EDIT: of course programming software, being something that uses a language, is actually a good candidate for LLM use and maybe teams can use LLMs to develop software more quickly to evaluate strategies and such.