Enabling on-device inference with up to 2 billion (2B) parameters, accelerating expansion into ultra-low-power edge AI ...