Sure - and the people responsible for a new freaking era of computing are the ones who asked "given how incredible it is that this works at all at 0.5b params, let's scale it up*.
It's not hyperbole - that it's an accurate description at a small scale was the core insight that enabled the large scale.
This hyperbole would describe any LLM of any size and quality, including a 0.5b model.