Arm is accelerating the move towards ubiquitous on-device artificial intelligence (AI) with a series of innovations unveiled in October 2025 that are set to redefine the landscape of edge intelligence. These developments cover a broad spectrum of applications—from gaming and neural graphics to embedded systems—illustrating a strategic emphasis on localising AI processing closer to users and data sources. The overarching goal is to enable more responsive, private, and efficient AI experiences without reliance on cloud connectivity, a shift that is poised to transform how intelligent experiences are created and delivered.
At the heart of this push is the general availability of ExecuTorch 1.0, an open-source inference engine co-developed with Meta. Designed to operate efficiently across Arm CPUs, GPUs, and NPUs, ExecuTorch leverages specialised hardware such as Scalable Matrix Extension 2 (SME2) for advanced matrix processing, Arm Neural Technology which scales inference, and the Ethos-U NPU family. This combination allows developers to deploy AI models with low latency and power consumption across a diverse array of edge devices—from minimalist microcontrollers to sophisticated smartphones. Industry experts highlight this as a critical milestone in making advanced AI accessible at the edge, significantly reducing dependence on cloud infrastructure while enhancing AI performance and privacy.
One of the most striking applications showcased is the capability to run large language models (LLMs) directly on smartphones, enhancing mobile gaming experiences with real-time conversational AI. In collaboration with Unreal Engine's Neural Network Engine and KleidiAI optimisations, Arm demonstrated this through the “Space Bartender” demo—a game featuring an interactive NPC powered by the Whisper voice model and a 4-bit quantised SmolLM2. This setup delivers a 2.3 times performance improvement while maintaining low power use, enabling network-free, immersive AI interactions on mobile devices. This breakthrough signals a paradigm shift for mobile gaming, where AI-driven characters can offer highly personalised and dynamic interactions entirely on-device, improving both user experience and data security.
Beyond conversational AI, Arm is driving innovation in visual computing and embedded intelligence through tools such as the Model Gym developer toolkit and the Arm Neural Graphics SDK. These enable developers to train and deploy neural graphics models like neural super-sampling (NSS) that enhance visual fidelity on mobile and embedded platforms, all while reducing computational and power costs. The Neural Graphics SDK includes Unreal Engine 5.4 integration for AI-powered rendering, promising near-console-level graphics efficiency on mobile devices. Such advancements empower game developers and device makers to deliver superior graphical performance in increasingly compact and energy-efficient ways.
The impact of these on-device AI innovations extends into real-world consumer products and specialised applications. For instance, the vivo X300 smartphone series incorporates SME2 technology to accelerate diverse AI workloads involving vision, speech, and text, demonstrating practical consumer integration. Meanwhile, Alif Semiconductor is advancing intelligent edge computing through embedded controllers powered by the Ethos-U85 NPU, enabling high-performance, low-power generative AI models tailored for smart homes, healthcare devices, and wearables. Even mobile photography benefits, with real-time neural-camera denoising technology on Arm-based devices using SME2, producing cleaner images and smoother 4K video in difficult low-light conditions, thus enhancing mobile camera capabilities.
Arm’s broad strategy integrates substantial software and hardware innovations supporting scalable AI performance at the edge. The integration of KleidiAI into ExecuTorch and other major AI frameworks such as Google's XNNPACK and MediaPipe further enhances performance without requiring developers to alter their codebases. Since 2024, KleidiAI has brought significant speed-ups and power efficiency gains to generative AI workloads on existing devices, including several-year-old smartphones and popular edge platforms like the Raspberry Pi 5. Coupled with advancements like SME2 and the Armv9 edge AI platform featuring the Cortex-A320 CPU and Ethos-U85 NPU, Arm’s ecosystem provides an 8-fold machine learning performance increase over previous generations, endorsed by industry leaders including AWS, Siemens, and Renesas.
This comprehensive on-device AI push underscores Arm’s foundational role in the next wave of intelligent computing. Through optimised hardware, open-source software, and tailored developer tools, Arm is not only enabling AI at the edge but making it practical and pervasive. The localisation of AI—from sophisticated language models to advanced neural graphics—promises enhanced user experiences, improved privacy, and a new range of possibilities across technology sectors. As AI moves decisively onto devices, Arm’s innovations lay critical groundwork for its widespread adoption and the future of edge intelligence.
📌 Reference Map:
- [1] (StartupHub) - Paragraphs 1, 2, 3, 4, 5, 6, 7
- [2] (Arm Newsroom) - Paragraphs 1, 2
- [3] (Arm Newsroom) - Paragraphs 5, 6
- [4] (Arm Newsroom) - Paragraph 6
- [5] (Arm Newsroom) - Paragraph 6
- [6] (Arm Newsroom) - Paragraph 3
- [7] (Arm Newsroom) - Paragraph 6
Source: Noah Wire Services