It is nothing new and has been well established in the literature since the 90s.
The shared article really is not worth the read and mostly uncovers an author who does not know what he write about.
LLMs didn’t exist in then. Attention only came out in 2017…
The network itself can be trained to solve most functions (or all, I forget precisely if NNs can solve all functions)
But the language model is not necessarily capable of solving all functions, because it was already trained on language.
It is nothing new and has been well established in the literature since the 90s.
The shared article really is not worth the read and mostly uncovers an author who does not know what he write about.