The thing that brought me to LLMs 3 years ago, was the ability to obtain custom-fit knowledge based on my context, avoiding the pathetic signal-to-noise ratio that the search engines bring.
The main focus now even with the huge models, is to make them as agentic as possible, and I can't help but think that, with the limited number of params, focusing on agentic task will surely degrade model's performance on other tasks.
Are there any LLM labs focusing on training a simple stupid model that has as much knowledge as possible? Basically an offline omniscient wikipedia alternative?
[link] [comments]




