Qualcomm's new AI Engine is designed to comprehensively utilize the AI capabilities of various cores within the SoC, such as CPU, NPU, and DSP.
First of all, the Hexagon NPU has been upgraded to be 37% faster than the previous generation, and it can run larger models with less power. In addition, the custom Oryon CPU core has the first-ever built-in matrix operation acceleration unit, allowing the CPU itself to quickly perform AI matrix calculations (e.g., multiplication-accumulation). This significantly reduces the latency of deep learning inference because the CPU, which is normally responsible for running apps, can directly assist with AI model inference.
By combining the advantages of heterogeneous cores according to the situation, it enables ultra-low latency (on-demand) responses for any AI task. As a result, within the smartphone,Always-on LLM (large language model)is realized, enabling a true agent-type AI assistant experience where the device continuously learns and understands the user and performs automatic tasks between apps according to the context.
For example, the phone AI learns the user's schedule, messages, location, etc., and turns on the meeting link in advance, or when a photo is taken, it instantly completes the editing with personal preference styles stored on the device. All of this personalized AI is processed internally on the device, so personal information is not sent externally, protecting privacy, and it works even in offline mode.
In addition, this AI Engine operates with improved power management, improving overall performance per watt and minimizing the impact of always-on AI tasks (e.g., always-listening triggers) on the battery.
In conclusion, the Qualcomm AI Engine is ahead of competitors in heterogeneous computing optimization and AI continuous operation capabilities, and is the key to realizing the AI experiences (generative AI, real-time translation subtitles, personal assistants, etc.) that were imagined on mobile.