Artificial Intelligence (AI) will see a significant shift out of the cloud and on to the edge (aka on-device, gateway, and on-premise server). This will happen initially in terms of inference (machine learning) and then by training. This shift means a huge opportunity for those chipset vendors with power-efficient chipsets and other products that can meet the demand for edge AI computing. Edge AI inference will grow from just 6% in 2017 to 43% in 2023, announced ABI Research, a market-foresight advisory firm providing strategic guidance on the most compelling transformative technologies.
“The shift to the edge for AI processing will be driven by cheaper edge hardware, mission-critical applications, a lack of reliable and cost-effective connectivity options, and a desire to avoid expensive cloud implementation. Consumer electronics, automotive, and machine vision vendors will play an initial critical role in driving the market for edge AI hardware. Scaling said hardware to a point where it becomes cost effective will enable a greater number of verticals to begin moving processing out of the cloud and on to the edge,” says Jack Vernon, Industry Analyst at ABI Research.
ABI Research has identified 11 verticals ripe for the adoption of AI, including automotive, mobile devices, wearables, smart home, robotics, small unmanned aerial vehicles, smart manufacturing, smart retail, smart video, smart building, and oil and gas sectors and split across a further 58 use cases. By 2023 the market will witness 1.2 billion shipments of devices capable of on-device AI inference – up from 79 million in 2017.
Cloud providers will still play a pivotal role, particularly when it comes to AI training. Out of the 3 billion AI device shipments that will take place in 2023, over 2.2 billion will rely on cloud service providers for AI training – this is still a real-term decline in the cloud providers market share for AI training, which currently stands around 99%, but will fall to 76% by 2023. Hardware providers should not be too concerned about this shift away from the cloud, as AI training is likely to be supported by the same hardware, only at the edge, either on-premise servers or gateway systems.
The power-efficient chipset is the main driver of edge AI. Mobile vendor Huawei is already introducing on-device AI training for battery power management in its P20 pro handset, in partnership with Cambricon Technologies. Chip vendors NVIDIA, Intel, and Qualcomm are also making a push to deliver the hardware that will enable automotive OEMs to experiment with on-device AI training to support their efforts in autonomous driving. Training at the edge on-device is beginning to gain momentum in terms of R&D, but it could still take some take some time for it to be a realist approach in most segments.
“The massive growth in devices using AI is positive for all players in the ecosystem concerned, but critically those players enabling AI at the edge are going to see an increase in demand that the industry to date has overlooked. Vendors can no longer go on ignoring the potential of AI at the edge. As the market momentum continues to swing toward ultra-low latency and more robust analytics, end users must start to incorporate edge AI in their roadmap. They need to start thinking about new business models like end-to-end integration or chipset as a service,” Vernon concludes.
These findings are from ABI Research’s Artificial Intelligence and Machine Learning market data. This report is part of the company’s AI and Machine Learning research service, which includes research, data, and Executive Foresights.