In reality, we shouldn’t be putting so much focus on the language model size and trying to define the next stage of artificial intelligence development by any significant measure of this kind.
Taking this to the extreme, while large language models (LLMs) like GPT are running out of data to train on and having difficulty scaling up, [DaveBben] is experimenting with scaling down instead ...
Large language models work well because they’re so large. The latest models from OpenAI, Meta and DeepSeek use hundreds of billions of “parameters” — the adjustable knobs that determine connections ...