前往小程序,Get更优阅读体验!
立即前往
首页
学习
活动
专区
工具
TVP
发布
社区首页 >专栏 >Cloud AI, Edge AI, Endpoint AI. What’s the Difference?

Cloud AI, Edge AI, Endpoint AI. What’s the Difference?

作者头像
用户6026865
发布2022-09-02 15:16:54
3450
发布2022-09-02 15:16:54
举报

Arm technology enables AI throughout the compute spectrum. Here, we explain the benefits and limitations of AI compute in cloud, edge and endpoint devices.

Today, the majority of what we call artificial intelligence (AI) is machine learning (ML), a subset of AI that involves machines learning from sets of data. In general, the greater the amount of data to learn from, the more the AI is able to infer meaning and the more useful it becomes.

Hundreds of thousands of gigabytes of data are generated every day in AI applications ranging from consumer devices to healthcare, logistics to smart manufacturing. With this much data generated, the key consideration is where that data should be processed.

Arm defines three categories within the compute spectrum: cloud, edge and endpoint. We can employ ML to process data within each of these but choosing the most suitable category isn’t as simple as going where the most compute performance is—as performance is only one governing factor in ensuring the learnings inferred from data remain useful.

What is Cloud AI?

Cloud AI refers to AI processing within powerful cloud data centers. For a long time, cloud AI was the obvious choice of compute platform to crunch enormous amounts of data. Were it not for the concept of shunting data from the edge and endpoint into cloud servers for hyper-efficient processing, AI would not be at the stage of maturity it enjoys today.

It’s likely the majority of AI heavy lifting will always be performed in the cloud due to its reliability, cost-effectiveness and concentration of compute—especially when it comes to training machine learning (ML) algorithms on historic data that doesn’t require an urgent response. Many consumer smart devices rely on the cloud for their ‘intelligence’: for example, today’s smart speakers give the illusion of on-device intelligence yet the only on-device AI they are capable of is to listen out for the trigger word (‘keyword spotting’).

Cloud AI is undisputed in its ability to solve complex problems using ML. Yet as ML’s use cases grow to include many mission-critical, real-time applications, these systems will live or die on how quickly decisions can be made. And when data has to travel thousands of miles from device to data center, there’s no guarantee that by the time it has been received, computed and responded to it will still be useful.

Applications such as safety-critical automation, vehicle autonomy, medical imaging and manufacturing all demand a near-instant response to data that’s mere milliseconds old. The latency introduced in asking the cloud to process that weight of data would in many cases reduce its value to zero.

It’s for this reason that many companies are now looking past the cloud to processing AI elsewhere in the compute infrastructure, moving compute nearer the data.

In a world where data’s time to value or irrelevancy may be measured in milliseconds, the latency introduced in transferring data to the cloud threatens to undermine many of the Internet of Things (IoT’s) most compelling use cases.

Edge AI moves AI and ML processing from the cloud to powerful servers at the edge of the network such as offices, 5G base stations and other physical locations very near to their connected endpoint devices. By moving AI compute closer to the data, we eliminate latency and ensure that all of that data’s value is retained.

Basic devices such as network bridges and switches have given way to powerful edge servers that add data center-level hardware into the gateway between endpoint and cloud. These powerful new AI-enabled edge servers, driven by new platforms such as Arm Neoverse, are designed to increase compute while decreasing power consumption, creating massive opportunities to instrument our cities, factories, farms, and environment to improve efficiency, safety, and productivity.

Edge AI has the potential to benefit both the data and the network infrastructure itself. At a network level, it could be used to analyze the flow of data for network prediction and network function management, while enabling edge AI to make decisions over the data itself offers significantly reduced backhaul to the cloud, negligible latency and improved security, reliability and efficiency across the board.

Another key function of edge AI is sensor fusion: combining the data from multiple sensors to create complex pictures of a process, environment or situation. Consider an edge AI device in an industrial application, tasked with combining data from multiple sensors within a factory to predict when mechanical failure might occur. This edge AI device must learn the interplay between each sensor and how one might affect the other and apply this learning in real-time.

There’s also a key security and resilience benefit in moving sensitive data no further than the edge: The more data we move to a centralized location, the more opportunities arise for that data’s integrity to be compromised. As the nature of compute changes, the edge is playing an increasingly crucial role in supporting diverse systems with a range of power and performance requirements. To deliver on service level agreements at scale for enterprises, the edge must embrace cloud-native software principles.

Arm is enabling this through Project Cassini, an open, collaborative, standards-based initiative to deliver a cloud-native software experience across a secure Arm edge ecosystem.

What is Endpoint AI?

Arm defines endpoint devices as physical devices connected to the network edge, from sensors to smartphones and beyond. As so much data is generated at the endpoint, we can maximise the insight we gain from that data by empowering endpoint devices to think for themselves and process what they collect without moving that data anywhere.

Due to their powerful internal hardware, smartphones have long been a fertile test-bed for endpoint AI. A smartphone camera is a prime example: it’s gone from something that takes grainy selfies to being secure enough for biometric authentication and powerful enough for computational photography – adding background blur (or a pair of bunny ears) to selfies in real-time.

This technology is now finding its way into smaller IoT devices. You may hear it referred to as the ‘AIoT’. In February 2020, Arm announced its solution for adding AI into even the smallest Arm-powered IoT devices. The Arm Cortex-M55 CPU and Arm Ethos-U55 micro neural processing unit (microNPU) combine to boost the performance of Arm-based Internet of Things (IoT) solutions by nearly 500 times—while retaining the trademark energy-efficient, cost-effective benefits our technology is known for. This technology will help to bring the benefits of Arm-powered compute to the IoT’s most challenging environments.

TinyML is an emerging sub-field of Endpoint AI, or AIoT, that enables ML processing in some of the very smallest endpoint devices containing microcontrollers no bigger than a grain of rice and consuming mere milliwatts of power.

Of course, endpoint AI also has its limitations: these devices are far more constrained in terms of performance, power and storage than edge AI and cloud AI devices. Data collected by one endpoint AI sensor can also have limited value on its own, as without the ‘top-down’ view of other data streams that sensor fusion at the edge enables, it is harder to see the full picture.

A combined, secure approach

Cloud AI, Edge AI and Endpoint AI each have their strengths and limitations. Arm’s range of heterogeneous compute IP scales the complete compute spectrum, ensuring that whatever your AI workload, Arm has a solution to enable it to be processed efficiently by putting intelligent compute power where it makes the most sense.

Most importantly, Arm technology ensures that data used in AI processing remains secure, from cloud to edge to endpoint. The Arm Platform Security Architecture (PSA) provides a platform, based on industry best-practice, that enables security to be consistently designed in at both a hardware and firmware level, while PSA Certified assures device manufacturers that their IoT devices are built secure. Within Arm processors, Arm TrustZone security technology simplifies IoT security and offers the ideal platform on which to build a device that adheres to PSA principles.

本文参与 腾讯云自媒体分享计划,分享自微信公众号。
原始发表:2022-06-14,如有侵权请联系 cloudcommunity@tencent.com 删除

本文分享自 SmellLikeAISpirit 微信公众号,前往查看

如有侵权,请联系 cloudcommunity@tencent.com 删除。

本文参与 腾讯云自媒体分享计划  ,欢迎热爱写作的你一起参与!

评论
登录后参与评论
0 条评论
热度
最新
推荐阅读
目录
  • Arm technology enables AI throughout the compute spectrum. Here, we explain the benefits and limitations of AI compute in cloud, edge and endpoint devices.
    • What is Cloud AI?
      • What is Endpoint AI?
        • A combined, secure approach
        相关产品与服务
        物联网
        腾讯连连是腾讯云物联网全新商业品牌,它涵盖一站式物联网平台 IoT Explorer,连连官方微信小程序和配套的小程序 SDK、插件和开源 App,并整合腾讯云内优势产品能力,如大数据、音视频、AI等。同时,它打通腾讯系 C 端内容资源,如QQ音乐、微信支付、微保、微众银行、医疗健康等生态应用入口。提供覆盖“云-管-边-端”的物联网基础设施,面向“消费物联”和 “产业物联”两大赛道提供全方位的物联网产品和解决方案,助力企业高效实现数字化转型。
        领券
        问题归档专栏文章快讯文章归档关键词归档开发者手册归档开发者手册 Section 归档