Dataset Viewer
Auto-converted to Parquet Duplicate
authors
sequencelengths
1
41
survey_title
stringlengths
24
120
year
stringdate
2019-01-01 00:00:00
2023-01-01 00:00:00
date
timestamp[ns, tz=UTC]date
2019-01-03 03:20:55
2023-12-18 07:47:33
category
stringclasses
23 values
abstract
stringlengths
0
2.64k
structure
listlengths
6
212
survey_id
int64
0
204
all_cites
sequencelengths
0
359
Bertopic_CD
float64
0.36
2.57
[ "Fang Liu", "Guoming Tang", "Youhuizi Li", "Zhiping Cai", "Xingzhou Zhang", "Tongqing Zhou" ]
A Survey on Edge Computing Systems and Tools
2019
2019-11-07T08:16:40
cs.DC
Driven by the visions of Internet of Things and 5G communications, the edge computing systems integrate computing, storage and network resources at the edge of the network to provide computing infrastructure, enabling developers to quickly develop and deploy edge applications. Nowadays the edge computing systems have received widespread attention in both industry and academia. To explore new research opportunities and assist users in selecting suitable edge computing systems for specific applications, this survey paper provides a comprehensive overview of the existing edge computing systems and introduces representative projects. A comparison of open source tools is presented according to their applicability. Finally, we highlight energy efficiency and deep learning optimization of edge computing systems. Open issues for analyzing and designing an edge computing system are also studied in this survey.
[ { "cite_extract_rate": 0, "cites": [], "content": "", "id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "level": "root", "origin_cites_number": 0, "parent_id": null, "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ] ], "subsections": [ "e3477cb3-f2f8-4d37-bfaf-3408b8c60cb3", "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "b72e7b0d-2142-415a-8bed-1b1f88160a2d", "e695a48d-424b-476b-aca3-58d54c720c98", "a3a2618a-548a-4cd2-95bc-06a60c44ffb0", "478327fb-c0c5-49d5-ae91-7fa642949abc", "aa0b39df-82ae-4dab-a3c4-17e99ed6d2be", "df3878a1-598b-4a96-97d4-ce6abf0afee7" ], "title": "root" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{sec:intro}\nIn the post-Cloud era, the proliferation of Internet of Things (IoT) and the popularization of 4G/5G, gradually changes the public's habit of accessing and processing data, and challenges the linearly increasing capability of cloud computing. Edge computing is a new computing paradigm with data processed at the edge of the network. Promoted by the fast growing demand and interest in this area, the edge computing systems and tools are blooming, even though some of them may not be popularly used right now. \nThere are many classification perspectives to distinguish different edge computing system. To figure out why edge computing appears as well as its necessity, we pay more attention to the basic motivations. Specifically, based on different design demands, existing edge computing systems can roughly be classified into three categories, together yielding innovations on system architecture, programming models and various applications, as shown in Fig.~\\ref{fig:systemCategory}. \n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.42\\textwidth]{systemsCategoryAll.pdf}\n\\caption{Categorization of edge computing systems.}\\label{fig:systemCategory}\n\\end{center}\n\\end{figure}\n\\begin{itemize}\n\\item Push from cloud. In this category, cloud providers push services and computation to the edge in order to leverage locality, reduce response time and improve user experience. Representative systems include Cloudlet, Cachier, AirBox, and CloudPath. Many traditional cloud computing service providers are actively pushing cloud services closer to users, shortening the distance between customers and cloud computing, so as not to lose market to mobile edge computing. For example, Microsoft launched AzureStack in 2017, which allows cloud computing capabilities to be integrated into the terminal, and data can be processed and analyzed on the terminal device.\n\\item Pull from IoT. Internet of Things (IoT) applications pull services and computation from the faraway cloud to the near edge to handle the huge amount of data generated by IoT devices. Representative systems include PCloud, ParaDrop, FocusStack and SpanEdge. Advances in embedded Systems-on-a-Chip (SoCs) have given rise to many IoT devices that are powerful enough to run embedded operating systems and complex algorithms. Many manufacturers integrate machine learning and even deep learning capabilities into IoT devices. Utilizing edge computing systems and tools, IoT devices can effectively share computing, storage, and network resources while maintaining a certain degree of independence.\n\\item Hybrid cloud-edge analytics. The integration of advantages of cloud and edge provides a solution to facilitate both global optimal results and minimum response time in modern advanced services and applications. Representative systems include Firework and Cloud-Sea Computing Systems\\nop{ and AWS Greengrass}. Such edge computing systems utilize the processing power of IoT devices to filter, pre-process, and aggregate IoT data, while employing the power and flexibility of cloud services to run complex analytics on those data. For example, Alibaba Cloud launched its first IoT edge computing product, LinkEdge, in 2018, which expands its advantages in cloud computing, big data and artificial intelligence to the edge to build a cloud/edge integrated collaborative computing system; Amazon released AWS Greengrass in 2017, which can extend AWS seamlessly to devices so that devices can perform local operations on the data they generate, while data is transferred to the cloud for management, analysis, and storage.\n\\end{itemize}\nFrom a research point of view, this paper gives a detailed introduction to the distinctive ideas and model abstractions of the aforementioned edge computing systems. Noted that the three categories are presented to clearly explain the necessity of edge computing, and the classification is not the main line of this paper. Specifically, we review systems designed for architecture innovation first, then introduce those for programming models and applications (in Sec.~\\ref{sec:edge-computing-systems}). Besides, some recently efforts for specific application scenarios are also studied.\n\\nop{This paper is from a research point of view, focusing on distinctive ideas and model abstractions of different systems. It introduces representative systems in the categories discussed above, and some recently proposed edge systems are outlined.}\nWhile we can find a lot of systems using edge computing as the building block, there still lacks standardization to such a paradigm. Therefore, a comprehensive and coordinated set of foundational open-source systems/tools are also needed to accelerate the deployment of IoT and edge computing solutions. Some open source edge computing projects have been launched recently (e.g., CORD). As shown in Fig.~\\ref{fig:systemCategory}, these systems can support the design of both architecture and programming models with useful APIs. We review these open source systems with a comparative study on their characteristics (in Sec.~\\ref{sec:open-source-projects}).\n\\nop{This paper studies current work on open source edge computing tools, which is challenging and has attracted a lot of attentions.}\nWhen designing the edge computing system described above, energy efficiency is always considered as one of the major concerns as the edge hardware is energy-restriction. Meanwhile, the increasing number of IoT devices is bringing the growth of energy-hungry services. Therefore, we also review the energy efficiency enhancing mechanisms adopted by the state-of-the-art edge computing systems from the three-layer paradigm of edge computing (in Sec.~\\ref{sec:energy-efficiency}).\n\\nop{This paper reviews the energy efficiency enhancing mechanisms adopted by the state-of-the-art edge computing systems from three Layer of the edge computing paradigm (cloud, edge server and device).}\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.42\\textwidth]{framework-intro.pdf}\n\\caption{Major building blocks and organization of this survey paper.}\\label{fig:framework-intro}\n\\vspace{-0.20in}\n\\end{center}\n\\end{figure}\nIn addition to investigations from the system view, we also look into the emerging techniques in edge computing system from the application view. Recently, deep learning based Artificial Intelligence (AI) applications are widely used and offloading the AI functions from the cloud to the edge is becoming a trend. However, deep learning models are known for being large and computationally expensive. Traditionally, many systems and tools are designed to run deep learning models efficiently on the cloud. As the multi-layer structure of deep learning, it is appropriate for edge computing paradigm, and more of its functions can be offloaded to the edge. Accordingly, this paper also studies the new techniques recently proposed to support the deep learning models at the edge (in Sec.~\\ref{sec:deep-learning-opt}).\n\\nop{This paper studies new techniques recently proposed to support the deep learning models at the edge; these techniques are categorized into three types: systems and toolkits, deep learning packages, and hardware.}\nOur main contributions in this work are as follows:\n\\begin{itemize}\n\\item Reviewing existing systems and open source projects for edge computing by categorizing them from their design demands and innovations. We study the targets, architecture, characteristics, and limitations of the systems in a comparative way.\n\\item Investigating the energy efficiency enhancing mechanism for edge computing from the view of the cloud, the edge servers, and the battery-powered devices.\n\\item Studying the technological innovations dedicated to deploying deep learning models on the edge, including systems and toolkits, packages, and hardware.\n\\item Identifying challenges and open research issues of edge computing systems, such as mobility support, multi-user fairness, and privacy protection.\n\\end{itemize}\nWe hope this effort will inspire further research on edge computing systems. The contents and their organization in this paper are shown in Fig.~\\ref{fig:framework-intro}. Besides the major four building blocks (Sec.~\\ref{sec:edge-computing-systems}$\\sim$Sec.~\\ref{sec:deep-learning-opt}), we also give a list of open issues for analyzing and designing an edge computing system in Sec.~\\ref{sec:key-design-issues}. The paper is concluded in Sec.~\\ref{sec:conclusion}.\n\\nop{The remaining parts of this paper are organized as follows. Section II introduces ten representative edge computing systems in academia, and outline some recently proposed edge systems. Section III presents some open source edge computing projects. Section IV discusses the energy efficiency enhancing mechanisms of edge computing systems. Section V presents new techniques that support the deep learning models at the edge. Open issues for analyzing and designing an edge computing system are studied in section VI. Finally, the paper is concluded in section VII.}", "id": "e3477cb3-f2f8-4d37-bfaf-3408b8c60cb3", "level": "section", "origin_cites_number": 0, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Introduction" ] ], "subsections": [], "title": "Introduction" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{sec:edge-computing-systems}\nIn this section, we review edge computing systems and tools presenting architecture innovations, programming models, and applications, respectively. For each part, we introduce work under the ``push\", ``pull\", and ``hybrid\" demand in turn.", "id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "level": "section", "origin_cites_number": 0, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ] ], "subsections": [ "41515cac-6ba7-49e0-b96b-0ef0d264a0ca", "de742ea0-6e5a-4f78-8cc0-f5db79790794", "44b0c252-9c6d-4438-9d77-129644ea6df0", "26e36cce-0199-4af9-92e8-65f39c1250d6", "13fec1ef-1af3-4617-8a90-614c3d31824e", "94a01773-4152-4ca2-a525-839a27b31c3b", "d41a4e11-d85b-40f4-b891-834279d27134", "87078893-bde3-463c-947c-f7dc4133bca0", "111f14e4-0ffb-419b-80b8-09b438ff1935", "49d0417a-3b91-46c8-8267-e0938c8f0aa3", "22a62c69-a6b7-4429-8620-50170782843d" ], "title": "Edge Computing Systems and Tools" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{Cloudlet.png}\n\\caption{Cloudlet Component Overview and Functions that support application mobility. \nA: Cloudlet Discovery, B: VM Provisioning, C: VM Handoff.}\\label{fig:Cloudlet}\n\\end{center}\n\\end{figure}\nIn 2009, Carnegie Mellon University proposed the concept of Cloudlet~, and the Open Edge computing initiative was also evolved from the Cloudlet project~. Cloudlet is a trusted, resource-rich computer or cluster of computers that are well-connected to the Internet and available to nearby mobile devices. It upgrades the original two-tier architecture ``Mobile Device-Cloud'' of the mobile cloud computing to a three-tier architecture ``Mobile Device-Cloudlet-Cloud''. Meanwhile, Cloudlet can also serve users like an independent cloud, making it a ``small cloud'' or ``data center in a box''. Although the Cloudlet project is not proposed and launched in the name of edge computing, its architecture and ideas fit those of the edge computing and thus can be regarded as an edge computing system. \nThe Cloudlet is in the middle layer of the three-tier edge computing architecture and can be implemented on a personal computer, low-cost server, or small cluster. {It can be composed of a single machine or small clusters consisting of multiple machines.} Like WiFi service access points, a Cloudlet can be deployed at a convenient location (such as a restaurant, a cafe, or a library). Multiple Cloudlets may form a distributed computing platform, which can further extend the available resources for mobile devices~. As the Cloudlet is just one hop away from the users' mobile devices, it improves the QoS with low communication delay and high bandwidth utilization.\nIn detail, Cloudlet has three main features as follows.", "id": "41515cac-6ba7-49e0-b96b-0ef0d264a0ca", "level": "subsection", "origin_cites_number": 3, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Cloudlet" ] ], "subsections": [ "52718c0a-d1e8-43e1-ab9c-697977382c63", "7177bd8d-5697-4a4d-b003-8e9ba628a902", "3937f91e-955f-4a8a-9d5c-45cfe802f164" ], "title": "Cloudlet" }, { "cite_extract_rate": 0, "cites": [], "content": "Cloudlet can be regarded as a small cloud computing center located at the edge of the network. Therefore, as the server end of the application, the Cloudlet generally needs to maintain state information for interacting with the client. However, unlike Cloud, Cloudlet does not maintain long-term state information for interactions, but only temporarily caches some state information. This reduces much of the burden of Cloudlet as a lightweight cloud.", "id": "52718c0a-d1e8-43e1-ab9c-697977382c63", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "41515cac-6ba7-49e0-b96b-0ef0d264a0ca", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Cloudlet" ], [ "subsubsection", "Soft State" ] ], "subsections": [], "title": "Soft State" }, { "cite_extract_rate": 0, "cites": [], "content": "Cloudlet has sufficient computing resources to enable multiple mobile users to offload computing tasks to it. Besides, Cloudlet also has stable power supply so it doesn't need to worry about energy exhaustion.", "id": "7177bd8d-5697-4a4d-b003-8e9ba628a902", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "41515cac-6ba7-49e0-b96b-0ef0d264a0ca", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Cloudlet" ], [ "subsubsection", "Rich Resources" ] ], "subsections": [], "title": "Rich Resources" }, { "cite_extract_rate": 0, "cites": [], "content": "Cloudlets are deployed at those places where both network distance and physical distance are short to the end user, making it easy to control the network bandwidth, delay and jitter. Besides, the physical proximity ensures that the Cloudlet and the user are in the same context (e.g., the same location), based on which customized services (e.g., the location-based service) could be provided.\nTo further promote Cloudlet, CMU built up an open edge computing alliance, with Intel, Huawei and other companies~, to develop standardized APIs for Cloudlet-based edge computing platforms. Currently, the alliance has transplanted OpenStack to the edge computing platform, which enables distributed Cloudlet control and management via the standard OpenStack APIs~. With recent development of the edge computing, the Cloudlet paradigm has been widely adopted in various applications, e.g., cognitive assistance system~, Internet of Things data analysis~, and hostile environments~. \nUnlike the cloud, cloudlets are deployed on the edge of the network and serve only nearby users. Cloudlet supports application mobility, allowing devices to switch service requests to the nearest cloudlet during the mobile process. As shown in Fig.~\\ref{fig:Cloudlet}, Cloudlet supports for application mobility relying on three key steps.\n\\textit{1) Cloudlet Discovery:} \nMobile devices can quickly discover the available Cloudlets around them, and choose the most suitable one to offload tasks.\n\\textit{2) VM Provisioning:}\nConfiguring and deploying the service VM that contains the server code on the cloudlet so that it is ready to be used by the client.\n\\textit{3) VM Handoff:}\nMigrating the VM running the application to another cloudlet.", "id": "3937f91e-955f-4a8a-9d5c-45cfe802f164", "level": "subsubsection", "origin_cites_number": 6, "parent_id": "41515cac-6ba7-49e0-b96b-0ef0d264a0ca", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Cloudlet" ], [ "subsubsection", "Close to Users" ] ], "subsections": [], "title": "Close to Users" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.4\\textwidth]{CloudPath.png}\n\\caption{CloudPath Architecture~.}\\label{fig:CloudPath}\n\\end{center}\n\\end{figure}\nCloudPath~ is an edge computing system proposed by the University of Toronto. In such a system, diverse resources like computing and storage are provided along the path from the user device to the cloud data center. It supports on-demand allocation and dynamic deployment of the multi-level architecture. The main idea of CloudPath is to implement the so-called ``path computing'', such that it can reduce the response time and improve the bandwidth utilization, compared with the conventional cloud computing. \nAs illustrated in Fig.~\\ref{fig:CloudPath}, the bottom layer of CloudPath is user devices, and the top layer is the cloud computing data center. The system reassigns those tasks of the data centers along the path (for path computing) to support different types of applications, such as IoT data aggregation, data caching services, and data processing services. Developers can select an optimal hierarchical deployment plan for their services by considering the factors such as cost, delay, resource availability and geographic coverage. Path computing builds a multi-tiered architecture, and from the top (traditional data center) to the bottom (user terminal equipment), the device capability becomes weaker, while the number of devices gets larger. On the premise of clear separation of computing and states, CloudPath extends the abstract shared storage layer to all data center nodes along the path, which reduces the complexity of third-party application development and deployment, and meanwhile keeps the RESTful development style.\nThe CloudPath application consists of a set of short-cycle and stateless functions that can be quickly instantiated at any level of the CloudPath framework. Developers either tag functions to specify where (such as edges, cores, clouds, etc.) their codes run, or tag performance requirements (such as response latency) to estimate the running location. \nCloudPath does not migrate a running function/module, but supports service mobility by stopping the current instance and re-starting a new one at the expected location.\nEach CloudPath node usually consists of the following six modules.\n\\textit{1) PathExecute:} implements a serverless cloud container architecture that supports lightweight stateless application functions/modules.\n\\textit{2) PathStore:} provides a distributed eventual consistent storage system that transparently manages application data across nodes. \\nop{PathStore is also used in PathDeploy and PathRoute to get the application program and routing information.}\n\\textit{3) PathRoute:} transmits the request to the most appropriate CloudPath node, according to the information such as user's location in the network, application preferences, or system status.\n\\textit{4) PathDeploy:} dynamically deploys and removes applications on CloudPath nodes based on application preferences and system policies.\n\\textit{5) PathMonitor:} provides real-time monitoring and historical data analysis function to applications and CloudPath nodes. It collects the metrics of other CloudPath modules on each node through the PathStore and presents the data to users using web pages.\n\\textit{6) PathInit:} is an initialization module in the top-level data center node and can be used to upload applications/services to the CloudPath.", "id": "de742ea0-6e5a-4f78-8cc0-f5db79790794", "level": "subsection", "origin_cites_number": 1, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "CloudPath" ] ], "subsections": [], "title": "CloudPath" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.4\\textwidth]{PCloud.pdf}\n\\caption{PCloud Architecture~.}\\label{fig:PCloud}\n\\end{center}\n\\end{figure}\nPCloud~ integrates the edge computing and storage resources with those at the cloud to support seamless mobile services. The architecture of PCloud is shown in Fig.~\\ref{fig:PCloud}. Specifically, these resources are virtualized through a special virtualization layer named STRATUS~, and form a distributed resource pool that can discover new resources and monitor resource changes. With the resource pool, the runtime mechanism is responsible for resource application and allocation. Through a resource description interface, the runtime mechanism selects and combines appropriate resources based on the requirements of specified applications. After the resources are combined, it generates a new instance to provide corresponding services for external applications, according to the resource access control policy. (Note that, the computing resources of the newly generated instance may come from multiple devices, which is equivalent to one integrated computing device for the external applications.) Thus, an application program is actually a combination of services running on the PCloud instance. For example, a media player application can be a combination of the storage, decoding and playing services. These services could be local or remote but are transparent to the applications. Furthermore, the PCloud system also provides basic system services, such as permission management and user data aggregation, to control the resources access of other users.\nIn the actual operation, the mobile application describes the required resources to the PCloud through interfaces. The PCloud will find out the optimal resource configuration by analyzing the description and the currently available resources, and then generates an instance to provide corresponding services for the application. \nPCloud integrates edge resources with cloud resources so that they can complement each other. The abundant resources from the cloud can make up for the lack of computing and storage capabilities at the edge; meanwhile, due to the physical proximity, edge devices can provide low-latency services to the user that cloud cannot offer. In addition, PCloud also enhances the availability of the entire system and can choose alternate resources when encountering network and equipment failures.", "id": "44b0c252-9c6d-4438-9d77-129644ea6df0", "level": "subsection", "origin_cites_number": 2, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "PCloud" ] ], "subsections": [], "title": "PCloud" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.45\\textwidth]{ParaDrop.png}\n\\caption{ParaDrop System~.}\\label{fig:ParaDrop}\n\\end{center}\n\\end{figure}\nParaDrop~ is developed by the WiNGS Laboratory at the University of Wisconsin-Madison. It is an edge computing framework that makes the computing/storage resources close to mobile devices and data sources available to the third party developers. Its goal is to bring intelligence to the network edge in a friendly way. \nParaDrop upgrades the existing access point to an edge computing system, which supports applications and services like a normal server. To isolate applications under the multi-tenancy scenario, ParaDrop leverages the lightweight container virtualization technique. As Fig.~\\ref{fig:ParaDrop} shows, the ParaDrop server (in the cloud) controls the deployment, starting and deletion of the applications. It provides a group of APIs, via which the developer can monitor and manage the system resources and configure the running environment. The web UI is also provided, through which the user can directly interact with the applications.\nThe design goals of ParaDrop include three aspects: multi-tenancy, efficient resource utilization and dynamic application management. To achieve these goals, the container technology is applied to manage the multi-tenancy resources separately.\\nop{Besides, the flexible APIs helps developers allocate specific resources to each application.} As the resources of the edge devices are very limited, compared with the virtual machine, the container consumes less resources and would be more suitable for delay sensitive and high I/O applications. Moreover, as applications running in the container, ParaDrop can easily control their startup and revocation.\nParaDrop is mainly used for IoT applications, especially IoT data analysis. Its advantages over traditional cloud system can be summarized as follows: a) since sensitive data can be processed locally, it protects the users’ privacy; b) WiFi access point is only one hop away from the data source, leading to low network delay and stable connection; c) only user requested data are transmitted to the equipment through the Internet, thus cutting down the total traffic amount and saving the bandwidth of the backbone network; d) the gateway can obtain the location information of the edge devices through radio signals (e.g., the distance between devices, and the location of the specific device), which facilitates the location-aware services; e) when edge devices cannot be connected to the Internet, the edge service can still work. \\nop{In a word, ParaDrop is well developed, the software system is fully open source, and the hardware devices that support ParaDrop system are also on the market.}", "id": "26e36cce-0199-4af9-92e8-65f39c1250d6", "level": "subsection", "origin_cites_number": 1, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "ParaDrop" ] ], "subsections": [], "title": "ParaDrop" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{SpanEdge.pdf}\n\\caption{SpanEdge Architecture~.}\\label{fig:SpanEdge}\n\\end{center}\n\\end{figure}\nStreaming processing is one important type of applications in edge computing, where data is generated by various data sources in different geographical locations and continuously transmitted in the form of streams. Traditionally, all raw data is transmitted over the WAN to the data center server, and stream processing systems, such as Apache Spark and Flink, are also designed and optimized for one centralized data center. However, this approach cannot effectively handle the huge data generated by a lot of devices at the edge of the network, and the situation is even worse when the applications require low latency and predictability. SpanEdge~ is a research project of the Royal Institute of Technology in Sweden. It unifies the cloud central node and the near-edge central node, reduces network latency in WAN connections, and provides a programming environment that allows the program to run near the data source. Developers can focus on developing streaming applications without considering where the data sources are located and distributed.\nThe data center in SpanEdge is composed of two levels: the cloud data center is the first level and the edge data center (such as the operator's cloud, Cloudlet, or Fog) is the second level. Partial streaming processing tasks run on the edge central nodes to reduce latency and boost performance. SpanEdge uses the master-worker architecture (as shown in Fig.~\\ref{fig:SpanEdge}) with one manager and multiple workers. The manager collects the streaming processing requests and assigns tasks to the workers. Workers mainly consist of cluster nodes whose primary responsibility is to execute tasks. There are two types of workers: hub-worker (first level) and spoke-worker (second level). \\nop{They have no functional differences.} The network transmission overhead and latency are related to the geographical location and network connection status between workers. The communication in SpanEdge is also divided into a system management communication (worker-manager) and a data transmission communication (worker-worker). System management communication aims to schedule tasks between managers and workers, and data transmission communication takes care of the data flow in each task. Each worker has an agent that handles system management operations, such as sending and receiving management information, monitoring compute nodes to ensure that they are running normally, and periodically sending heartbeat messages to the manager to ensure immediate recovery when the task fails.\nSpanEdge allows developers to divide the tasks into local ones and global ones. Local tasks should run on the node near the data source and provide only part of the required data; global tasks are responsible for further processing the results of local tasks and aggregating all results. SpanEdge creates a copy of the local task on each spoke-worker which has all corresponding data sources. If the data is insufficient, the task is dispatched to the hub-worker. The global task runs on a separate hub-worker, and the scheduler selects the optimal hub-worker based on the network delay (i.e. the distance from the spoke-worker in the network topology).", "id": "13fec1ef-1af3-4617-8a90-614c3d31824e", "level": "subsection", "origin_cites_number": 1, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "SpanEdge" ] ], "subsections": [], "title": "SpanEdge" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tp]\n\t\\centering\n\t\\includegraphics[width=0.9\\linewidth]{Cloud-Sea.pdf}\n\t\\caption{Cloud-Sea Computing Model.}\n\t\\label{fig:Cloud-Sea}\n\\end{figure}\nCloud-Sea Computing Systems project~ is a main research thrust of the Next Generation Information and Communication Technology initiative (the NICT initiative) and a 10-year strategic priority research initiative, launched by the Chinese Academy of Science in 2012. The NICT initiative aims to address the three major technology challenges in the coming Zettabyte era: i) improving the performance per watt by $1000$ times, ii) supporting more applications from the human-cyber-physical ternary computing, and iii) enabling transformative innovations in devices, systems and applications, while without polluting beneficial IT ecosystems.\nIn the cloud-sea computing system, ``cloud'' refers to the datacenters and ``sea'' refers to the terminal side (the client devices, e.g., human-facing and physical world facing subsystems). The design of the project can be depicted from three levels: the overall systems architecture level, the datacenter server and storage system level, and the processor chip level. The project contains four research components: a computing model called REST 2.0 which extends the representational state transfer (REST)~ architectural style of Web computing to cloud-sea computing, a three-tier storage system architecture capable of managing ZBs of data, a billion-thread datacenter server with high energy efficiency, and an elastic processor aiming at energy efficiency of one trillion operations per second per watt.\nAs shown in Fig.~\\ref{fig:Cloud-Sea}, the cloud-sea computing model includes sea-side functions and cloud-side functions. The sea zone is expanded from the traditional cloud client, e.g., a home, an office, a factory manufacturing pipeline. There can be multiple client devices inside a sea zone, and each device can be human facing or physical world facing. In a sea zone, there is a special device (like a home datacenter or a smart TV set) designated as the seaport for three purposes: i) a gateway interfacing the sea zone to the cloud, ii) a gathering point of information and functionalities inside a sea zone, and iii) a shield protecting security and privacy of the sea zone. A device inside a sea zone does not communicate to the cloud directly, but through the seaport, either implicitly or explicitly. The SeaHTTP, a variant based on HTTP $2.0$, is the widely-used protocol in sea zones of the cloud-sea system to connect with the cloud.\nThe cloud-sea computing model has four distinct features. \n\\textit{1) Ternary computing via sea devices:} \nHuman and physical world entities interface and collaborate with the cyberspace through the sea side. For example, users can leverage a smart phone application to read and control a sensor device in a home through an Internet application service.\n\\textit{2) Cooperation with locality:} \nA specific network computing system will partition its functions between the sea side and the cloud side. Sea-side functions include sensing, interaction, and local processing while cloud-side functions include aggregations, request-response, and big data processing.\n\\textit{3) Scalability to ZB and trillion devices:} \nThis future Net will collectively need to support trillions of sea devices and to handle ZBs of data.\n\\textit{4) Minimal extension to existing ecosystems:} \nThe REST $2.0$ cloud-sea computing architecture attempts to utilize existing Web computing ecosystems as much as possible. \nOverall, the cloud-sea computing model is proposed to migrate the cloud computing function to the sea side, and it focuses more on the devices at the ``sea'' side and the data at the ``cloud'' side. Typical edge computing is more generalized and may care about any intermediate computing resources and network resources between the ``sea'' and the ``cloud''. The researches in cloud-sea computing (e.g., energy efficient computing and elastic processor designing) are consultative for the edge computing.", "id": "94a01773-4152-4ca2-a525-839a27b31c3b", "level": "subsection", "origin_cites_number": 2, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Cloud-Sea Computing Systems" ] ], "subsections": [], "title": "Cloud-Sea Computing Systems" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{subsec:Cachier_and_Precog}\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{Cachier.pdf}\n\\caption{Cachier System~.}\\label{fig:Cachier}\n\\end{center}\n\\end{figure}\nCachier~ and Precog~ are two edge caching systems that proposed by the researchers from Carnegie Mellon University for image recognition. \nRecognition applications have strict response time requirements, while the computation is huge due to the model complexity and dataset size. With edge computing, we can leverage the computing resource of the edge nodes to process the matching, so the network delay can be reduced.\nConsidering that edge nodes mainly provide services to nearby users, the spatio-temporal characteristics of service requests can be leveraged. From the perspective of caching, Cachier system proposes that edge nodes can be used as ``computable cache devices'', which can cache recognition results, reduce matching dataset size and improve response time. By analyzing the response delay model and requests' characteristics, the system can dynamically adjust the size of the dataset on the edge nodes according to the environment factors, thus ensuring optimal response time. \nAs Fig.~\\ref{fig:Cachier} shows, Cachier consists of the Recognizer Module, the Optimizer Module, and the Offline Analysis Module. \nThe Recognizer Module is responsible for analyzing and matching the received figures according to the cached training data and model. If there is a match, the result will be directly returned to the user, otherwise the figure will be transmitted to the cloud for recognition. \nThe distribution estimator in the Optimizer Module can use the maximum a posteriori estimation to predict the request distribution. Secondly, given the classification algorithm and training dataset, the cache searching delay and the cache accuracy can also be calculated by the Offline Analysis Module. At last, the Profiler sub-module is responsible for estimating the network delay and cloud latency incurred by cache misses. It measures and records the delay under corresponding distance in real time, and uses the moving average filter to remove the noise data. By taking such information into the delay expectation time model, Cachier is able to calculate the optimal cache size, and then adjusts the cache on the edge nodes accordingly. \\nop{In addition, the Cachier system also provides other optimizations, such as hot start and feature request waiting queues.}\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{Precon.png}\n\\caption{Precog System~.}\\label{fig:Precon}\n\\end{center}\n\\end{figure}\nPrecog is an extension of Cachier. The cached data is not only on the edge nodes, but also on the end devices which are used for selection calculation in order to reduce the image migration between the edge nodes and the cloud. Based on the prediction model, Precog prefetches some of the trained classifiers, uses them to recognize the images, and cooperates with the edge nodes to efficiently complete the tasks. \nPrecog pushes computing capability to the end device and leverages the locality and selectivity of the user requests to reduce the last mile delay. For the end device, if the cache system uses the same cache replacement policy as the edge node applies, it will lead to a large number of forced misses. For a single user, the device only has the request information of the user, and usually the user will not identify the same object multiple times in the near future. Therefore, Precog constructs a Markov model using the request information from the edge node to describe the relationship among the identified objects, and then predicts the potential future requests. \nPrecog improves the delay expectation time model by considering the network state, device capabilities and the prediction information from edge nodes. \nThe system architecture of Precog is illustrated in Fig.~\\ref{fig:Precon}. We can see that it is mainly composed of the feature extraction module, the offline analysis module, the optimizer module and the profiler module. The edge node is mainly responsible for the construction of Markov model. Based on the offline analysis results, the Markov model prediction results, and the network information provided by the profiler module, the optimizer module determines the number of feature to be extracted as well as the data to be cached on the end device.", "id": "d41a4e11-d85b-40f4-b891-834279d27134", "level": "subsection", "origin_cites_number": 2, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Cachier and Precog" ] ], "subsections": [], "title": "Cachier and Precog" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{FocusStack.pdf}\n\\caption{FocusStack System~.}\\label{fig:FocusStack}\n\\end{center}\n\\end{figure}\nFocusStack~ is developed by AT\\&T Labs, which supports the deployment of complex applications on a variety of potential IoT edge devices. Although the resources, such as computing, power consumption, connectivity, are limited on edge devices, they have the nature of mobility. Hence, it is very important to have a system that can discover and organize available edge resources. FocusStack is such a system that can discover a set of edge devices with sufficient resources, deploy applications and run them accordingly. Thus, the developers can focus more on the design of applications than on how to find and track edge resources.\nFocusStack consists of two parts (as shown in Fig.~\\ref{fig:FocusStack}): i) Geocast system, which provides location-based situational awareness (LSA) information, and ii) OpenStack extension (OSE), which is responsible for deploying, executing, and managing the containers on the edge devices. FocusStack builds a hybrid cloud of edge devices (containers) and data center servers (virtual machines). When a user initiates a cloud operation through the FocusStack API (such as instantiating a container), the LSA subsystem analyzes the scope of the request based on the Geocast route and sends a resource list of geographic locations to the target area, and waits for the response from (online) edge devices that can satisfy the requirements. Then, the selected edge device runs the corresponding OpenStack operations with the help of the conductor module.\nThe situational aware subsystem enables one group of devices to monitor the survival status of each other, and each device can update the sensing information of other devices. \nThe geo-processing projection service is primarily intended to pass requests and reply messages between areas of interest, send out device control information (like drones) and broadcast location-based information. The service is based on a two-layer network, and data can be transmitted over a dedicated WiFi network or the Internet. The sender transmits the data packet to the geo-router server which tracks the location and metadata of each edge device, and then the data packet is sent to each device (including edge devices and a cloud device running a SAMonitor instance) in the specified area based on the address. Location and connection statuses are maintained by the geo-routing database (GRDB). The GCLib is a software framework that provides data acquisition services, and it runs on edge devices and cloud applications that use SAMonitor. \nAny application or service in need of the state of the device in the area needs a SAMonitor component to communicate with the LSA subsystem. The application server makes a request to SAMonitor through the FocusStack API, and then SAMonitor builds a real-time graph of the current area and returns a list of available edge devices. The regional graph is sent to the conductor in the OSE, which is responsible for checking whether the devices are capable of running the tasks, whether the pre-defined policy rules are met, and so on. After that, the available edge device list is submitted to the application server and the server selects devices to be used. OSE manages and deploys the program through the OpenStack Nova API. The edge devices run a custom version of Nova Compute to interact with the local Docker to manage the container. The container on the edge devices supports all OpenStack services, including access to virtual networks and application-based granularity configuration.", "id": "87078893-bde3-463c-947c-f7dc4133bca0", "level": "subsection", "origin_cites_number": 1, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "FocusStack" ] ], "subsections": [], "title": "FocusStack" }, { "cite_extract_rate": 0, "cites": [], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{AirBox.png}\n\\caption{AirBox Architecture~.}\\label{fig:AirBox}\n\\end{center}\n\\end{figure}\nAirBox~ is a secure, lightweight and flexible edge function system developed by Georgia Institute of Technology. It supports fast and flexible edge function loading and provides service security and user privacy protection. The Edge Function (EF) in AirBox is defined as a service that can be loaded on an edge node, and the software stack that supports the edge function is named Edge Function Platform (EFP). \nAs shown in Fig.~\\ref{fig:AirBox}, the AirBox consists of two parts: the AB console and the AB provisioner. The back-end service manager deploys and manages the EF on the edge nodes through the AB console. The AB provisioner which runs at the edge is responsible for providing dynamic EF seamlessly. The edge functions are implemented through system-level containers with minimal constraints on developers. \nSecurity is enhanced by using the hardware security mechanisms like Intel SGX. AirBox provides centrally controlled backend services in discovering edge nodes and registering them. The AB console is a web-based management system, which activates the docker start-up process on the edge node with AB provisioner. \nTo ensure the security of the AirBox, the EF consists of a trusted part and an untrusted part. The untrusted part is responsible for all network and storage interactions. Based on OpenSGX APIs, AirBox provides four extensible APIs to implement secure communication and storage: Remote Attestation, Remote Authentication, Sealed Storage, and EF Defined Interface. AirBox supports a variety of edge features such as aggregation, buffering, and caching.", "id": "111f14e4-0ffb-419b-80b8-09b438ff1935", "level": "subsection", "origin_cites_number": 1, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "AirBox" ] ], "subsections": [], "title": "AirBox" }, { "cite_extract_rate": 0, "cites": [], "content": "Wayne State University's MIST Lab proposes a programming model for edge computing\\textemdash Firework~. In the model of Firework, all services/functions are represented in a data view of datasets and functions which can be the results of processing with their own data or secondary processing with other services/datasets. A system consisting of nodes that apply the Firework model is called the Firework system. In the Firework system, instead of dividing nodes into edge nodes and cloud ones, they are all considered as Firework nodes. Through the mutual invocation of the Firework nodes, the data can be distributed and processed on each node. Along the path of data transmission, the edge nodes perform a series of calculations upon the data, thus forming a ``computational flow''. The beginning of the computational flow could be user terminals, edge servers close to the user, edge servers close to the cloud, or cloud nodes. \nIn the Firework system, the data processing service is split into multiple sub-services, and the scheduling of the data processing is performed at the following two layers.\n\\textit{1) Same sub-service layer scheduling:} \nA Firework node can cooperate with another for the same sub-services in the surrounding area to achieve optimal response time. For idle Firework nodes, the system can schedule sub-service programs onto them, such that a cluster providing specific sub-service is formed dynamically and complete the service faster.\n\\textit{2) Computational flow layer scheduling:} \nFirework nodes along the computational flow can cooperate with each other and dynamically schedule execution nodes to achieve an optimal solution. For example, depending on the states of the network, the system can choose Firework nodes for service providing based on the nodes' locations (e.g., selecting those closest to the users).\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.35\\textwidth]{Firework.jpg}\n\\caption{An example of a Firework instance that consists of heterogeneous computing platforms~.}\\label{fig:Firework}\n\\end{center}\n\\end{figure}\nAs shown in Fig.~\\ref{fig:Firework}, Firework divides the nodes into computing nodes and managers, depending on the type of service those nodes provide. In general, each node with the Firework model has three modules: the job management module, the actuator management module, and the service management module.\n\\begin{table*}[htb]\n\\centering\n\\caption{Summary of edge computing systems} \n\\label{tab:otherEdgeSystems}\n\\begin{tabular}{cccccc}\n\\hline\n\\textbf{\\begin{tabular}[c]{@{}c@{}}Application\\\\ Scenarios\\end{tabular}} & \\textbf{\\begin{tabular}[c]{@{}c@{}}Edge\\\\ Computing\\\\ Systems\\end{tabular}} & \\textbf{\\begin{tabular}[c]{@{}c@{}}End\\\\ Devices\\end{tabular}} & \\textbf{Edge Nodes} & \\textbf{\\begin{tabular}[c]{@{}c@{}}Computation\\\\ Architecture\\end{tabular}} & \\textbf{Features/Targets} \n \\\\ \\hline\n\\multirow{10}{*}{\\begin{tabular}[c]{@{}c@{}} General \\\\ Usage \\\\ Scenario \\end{tabular}} & Cloudlet & Mobile devices & Cloudlet & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Lightweight\\\\ VM migration\\end{tabular} \n\\\\ \\cline{2-6}\n & PCloud & Mobile devices & \\begin{tabular}[c]{@{}c@{}}Mobile devices,\\\\ local server, PC\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Resource integration, \\\\ dynamic allocation\\end{tabular} \n\\\\ \\cline{2-6}\n & ParaDrop & IoT devices & Home gateway & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Hardware, \\\\ developer support\\end{tabular}\n\\\\ \\cline{2-6}\n & Cachier \\& Precog & Mobile devices & \\begin{tabular}[c]{@{}c@{}}Mobile devices,\\\\ local server, PC\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Figure recognition, \\\\ identification \\end{tabular}\n\\\\ \\cline{2-6}\n & FocusStack & IoT devices & Router, server & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Location-based info, \\\\ OpenStack extension \\end{tabular}\n\\\\ \\cline{2-6}\n & SpanEdge & IoT devices & \\begin{tabular}[c]{@{}c@{}}Local cluster,\\\\ Cloudlet, Fog \\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (2-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Streaming processing, \\\\ local/global task \\end{tabular}\n\\\\ \\cline{2-6}\n & AirBox & IoT devices & \\begin{tabular}[c]{@{}c@{}}Mobile devices,\\\\ local server, PC\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & Security\n\\\\ \\cline{2-6}\n & CloudPath & Mobile devices & \n \\begin{tabular}[c]{@{}c@{}}Multi-level \\\\data centers\\end{tabular}& \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (multi-tier)\\end{tabular} & Path computing\n\\\\ \\cline{2-6}\n & Firework & Firework.Node & \n Firework.Node & \\begin{tabular}[c]{@{}c@{}}Two-layer\\\\ scheduling\\end{tabular} & Programming model\n\\\\ \\cline{2-6}\n & Cloud-Sea & Sea & \n Seaport & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Minimal extension,\\\\transparency\\end{tabular}\n\\\\ \\hline\n\\multirow{2}{*}{\\begin{tabular}[c]{@{}c@{}}Vehicular\\\\ Data\\\\ Analytics\\end{tabular}} & OpenVDAP & CAVs & XEdge & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (2-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}General\\\\ platform\\end{tabular} \\\\ \\cline{2-6}\n & SafeShareRide & \\begin{tabular}[c]{@{}c@{}}Smartphones\\\\ and vehicles\\end{tabular} & Smartphones & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (2-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}In-vehicle\\\\ security\\end{tabular} \\\\ \\hline\n\\multirow{2}{*}{\\begin{tabular}[c]{@{}c@{}}Smart\\\\ Home\\end{tabular}} & Vigilia & \\begin{tabular}[c]{@{}c@{}}Smart\\\\ home devices\\end{tabular} & Hubs & \\begin{tabular}[c]{@{}c@{}}Edge\\\\ only\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Smart\\\\ Home Security\\end{tabular} \\\\ \\cline{2-6}\n & HomePad & \\begin{tabular}[c]{@{}c@{}}Smart\\\\ home devices\\end{tabular} & Routers & \\begin{tabular}[c]{@{}c@{}}Edge\\\\ only\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Smart\\\\ Home Security\\end{tabular} \\\\ \\hline\n\\multirow{3}{*}{\\begin{tabular}[c]{@{}c@{}}Video\\\\ Stream\\\\ Analytics\\end{tabular}} & LAVEA & $\\sim$ & $\\sim$ & \\begin{tabular}[c]{@{}c@{}}Edge\\\\ or cloud\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Low\\\\ latency response\\end{tabular} \\\\ \\cline{2-6}\n & VideoEdge & Cameras & \\begin{tabular}[c]{@{}c@{}}Cameras and\\\\ private\\\\ clusters\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Hybrid\\\\ (3-tier)\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Resource-accuracy\\\\ tradeoff\\end{tabular} \\\\ \\cline{2-6}\n & \\begin{tabular}[c]{@{}c@{}}Video\\\\ on drones\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Autonomous\\\\ drones\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Portable\\\\ edge computers\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Edge\\\\ only\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Bandwidth\\\\ saving\\end{tabular} \\\\ \\hline\n\\begin{tabular}[c]{@{}c@{}}Virtual\\\\ Reality\\end{tabular} & MUVR & Smartphones & \\begin{tabular}[c]{@{}c@{}}Individual\\\\ households\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Edge\\\\ only\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Resource utilization\\\\ efficiency\\\\ optimization\\end{tabular} \\\\ \\hline\n\\end{tabular}\n\\end{table*}\n\\textit{1) Service management module:} \nThis type of module is designed for the management of data views. It provides interfaces to update the data view, as well as relevant programs for data processing.\n\\textit{2) Job management module:}\nThis type of module is responsible for the scheduling, monitoring, and evaluation of the task executions. When the local computing resources are insufficient, the module can look into the node list and the data view\\nop{ (of the Firework manager)} and make resource re-scheduling at the same sub-service layer. When the sub-service is running, the module can also provide necessary monitoring information and give feedback to other upstream and downstream nodes for flow layer scheduling.\n\\textit{3) Actuator management module:}\nThis type of module is mainly responsible for managing all hardware resources and hosting the execution processes of different tasks. With the help of this module, the device, running environment and the upper layer functions could be decoupled, so that the nodes of Firework system are not limited to a certain type of devices, and the data processing environment is not limited to a certain type of computing platform.\n\\nop{\nFirework manager has two sub-modules, and Firework computing node has three sub-modules. For Firework manager, it includes service management module that provides the registration and viewing capabilities of the view and the status of the associated Firework compute nodes, and job management module, which is responsible for scheduling the service. Firework computing node refers to a node that provides computing resources, which may have its own database and provide a corresponding view of the data but may also simply provide computing resources. \n}", "id": "49d0417a-3b91-46c8-8267-e0938c8f0aa3", "level": "subsection", "origin_cites_number": 1, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Firework" ] ], "subsections": [], "title": "Firework" }, { "cite_extract_rate": 0, "cites": [], "content": "The edge systems introduced above depict some typical and basic innovations on the exploitation of edge analytics for highly-responsive services. The previous systems are used in general cases, which lay the foundation of further development. We highlight that, in addition to these efforts, there are many other edge systems tuned for a series of different application scenarios.In Table~\\ref{tab:otherEdgeSystems}, we briefly summarize the general and application-specific edge systems, which leverage different kinds of edge nodes to serve diverse end devices using either a hybrid or edge only computation architecture.\nCompared to both on-board computation and cloud-based computation, edge computing can provide more effective data analytics with lower latency for the moving vehicles~. In~, an open full-stack edge computing-based platform OpenVDAP is proposed for the data analytics of connected and autonomous vehicles (CAVs). OpenVDAP proposes systematic mechanisms, including varied wireless interfaces, to utilize the heterogeneous computation resources of nearby CAVs, edge nodes, and the cloud. For optimal utilization of the resources, a dynamic scheduling interface is also provided to sense the status of available resources and to offload divided tasks in a distributed way for computation efficiency. SafeShareRide is an edge-based attack detection system addressing in-vehicle security for ridesharing services~. Its three detection stages leverage both the smartphones of drivers and passengers as edge computing platform to collect multimedia information in vehicles. Specifically, speech recognition and driving behavior detection stages are first carried out independently to capture in-vehicle danger, and video capture and uploading stage is activated when abnormal keywords or dangerous behaviors are detected to collect videos for cloud-based analysis. By using such an edge-cloud collaborative architecture, SafeShareRide can accurately detect attacks in-vehicle with low bandwidth demand.\nAnother scenario that edge computing can play an important role is the IoT devices management in the smart home environment. Wherein, the privacy issue of the wide range of home-devices is a popular topic. In~, the Vigilia system is proposed to harden smart home systems by restricting the network access of devices. A default access deny policy and an API-granularity device access mechanism for applications are adopted to enforce access at the network level. Run time checking implemented in the routers only permits those declared communications, thus helping users secure their home-devices. Similarly, the HomePad system in~ also proposes to execute IoT applications at the edge and introduces a privacy-aware hub to mitigate security concerns. Homepad allows users to specify privacy policy to regulate how applications access and process their data. Through enforcing applications to use explicit information flow, Homepad can use Prolog rules to verify whether applications have the ability to violate the defined privacy policy at install time.\nEdge computing has also been widely used in the analysis of video stream. LAVEA is an edge-based system built for latency-aware video analytics nearby the end users~. In order to minimize the response time, LAVEA formulates an optimization problem to determine which part of tasks to be offloaded to the edge computer and uses a task queue prioritizer to minimize the makespan. It also proposes several task placement schemes to enable the collaboration of nearby edge nodes, which can further reduce the overall task completion time. VideoEdge is a system that provides the most promising video analytics implementation across a hierarchy of clusters in the city environment~. A 3-tier computation architecture is considered with deployed cameras and private clusters as the edge and remote server as the cloud. The hierarchical edge architecture is also adopted in~ and is believed to be promising in processing live video stream at scale. Technically, VideoEdge searches thousands of combinations of computer vision components implementation, knobs, and placement and finds a configuration to balance the accuracy and resource demands using an efficient heuristic. In~, a video analytics system for autonomous drones is proposed, where edge computing is introduced to save the bandwidth. Portable edge computers are required here to support dynamic transportation during a mission. Totally four different video transmission strategies are presented to build an adaptive and efficient computer vision pipeline. In addition to the analytics work (e.g., object recognition), the edge nodes also train filters for the drones to avoid the uploading of the uninteresting video frames.\nIn order to provide flexible virtual reality (VR) on untethered smartphones, edge computing can be useful to transport the heavy workload from smartphones to their nearby edge cloud~. However, the rendering task of the panoramic VR frames (i.e., 2GB per second) will also saturate the individual households as common edge in the house. In~, the MUVR system is designed to support multi-user VR with efficient bandwidth and computation resources utilization. MUVR is built on a basic observation that the VR frames being rendered and transmitted to different users are highly redundant. For computation efficiency, MUVR maintains a two-level hierarchical cache for invariant background at the edge and the user end to reuse frames whenever necessary. Meanwhile, MUVR transmits a part of all frames in full and delivers the distinct portion for the rest frames to further reduce the transmission costs.\n\\nop{", "id": "22a62c69-a6b7-4429-8620-50170782843d", "level": "subsection", "origin_cites_number": 9, "parent_id": "2d1d8056-04f7-4532-9d27-b8a8f91e3db7", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Edge Computing Systems and Tools" ], [ "subsection", "Other Edge Computing Systems" ] ], "subsections": [], "title": "Other Edge Computing Systems" }, { "cite_extract_rate": 0, "cites": [], "content": "ParaDrop is an edge computing system that uses a Wi-Fi access point/wireless gateway as a hardware platform to provide third-party applications with computing and storage resources at the edge of the network. Compared to cloud data centers, ParaDrop is closer to data sources and end users. It is aware of network and environmental status, which helps reduce response time, enhance data security and protect user privacy.", "id": "b72e7b0d-2142-415a-8bed-1b1f88160a2d", "level": "section", "origin_cites_number": 0, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "ParaDrop" ] ], "subsections": [ "b08f7313-052a-475e-91fc-ecf9b1117462", "adac1e95-6fb5-4cca-b7a1-c16f16b2589a", "e332c996-588d-441c-bafd-1d99b5403e4d" ], "title": "ParaDrop" }, { "cite_extract_rate": 0, "cites": [], "content": "Since computing and storage resources on the gateway are available for most of the time, ParaDrop prefer to use gateways and Wi-Fi access points to deploy service which are named as chutes. ParaDrop provides developers with a full-featured API and centralized control framework that simplifies deployment and management complexity. \nParaDrop contains mainly three components: Cloud Controller, Edge Compute Node and APIs. Edge services/chutes are deployed on edge node (e.g. gateways) through the backend Cloud Controller and developer API. There can be multiple chutes on one edge node, the isolation is guaranteed by using container technology. In Figure Y, two services are used as examples to illustrate the usage scenarios. SecCam is a wireless security monitoring service that collects video data within the range and analyzes activities locally. EnvSense is an environmental monitoring service that monitors temperature and humidity in buildings.\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{ParaDrop_2.png}\n\\caption{ParaDrop System}\\label{fig:ParaDrop_2}\n\\end{center}\n\\end{figure}\nParaDrop Cloud Controller manages ParaDrop system resources, maintains information of edge compute nodes and users, and provides the Chute Store (like Google Play Store and AppStore) for software distribution. From the programming viewpoint, the ParaDrop Cloud Controller supports two important interfaces: the WAMP API and the HTTP RESTful API. The WAMP (Web Application Messaging Protocol) API is used to communicate with the edge compute nodes, send control information, and receive replies and status reports in real time. The HTTP RESTful API is used to communicate with users, developers and administrators. Specifically, the backend aggregates the data from all authenticated edge compute nodes to provide a graphical display to the users. From the functionality viewpoint, the ParaDrop Cloud Controller saves the deployment information of the ParaDrop system, such as the location and configuration of edge compute nodes; and transfers messages between the user and the edge compute node. Noted that because of the highly distributed nature of edge computing, the central cloud controller is not strictly required, each edge compute node has a publicly-documented local API and can be directly managed using command line tools.\nParaDrop Edge Compute Node is the specific computing platform that provides a virtualized resource environment for services/chutes, including CPU, memory and network resources. The software platform is entirely built from open source components, each chute is deployed as Docker containers so that the hardware heterogeneity is shield. A ParaDrop Daemon runs on the edge devices and responsible for: 1) registering the current edge device at the ParaDrop Cloud Controller ; 2) monitoring the status of the edge device and reporting it to the Cloud Controller; 3) managing local resources and processes, including virtual wireless devices, firewalls, DHCP, and so on; 4) managing the running containers and responsible for the installation, start, stop and uninstall of chutes based on the information received from the Cloud Controller.\nParaDrop API is used to exports the system capabilities to developers so that they can monitor and control the ParaDrop system. The API is composed of two parts: the cloud API and the edge API. The cloud API provides fixed status information (such as user type, user authorization, chute description, chute resource configuration, edge compute node configuration) and real-time status information (such as chute and edge compute node running status). Besides, the cloud API can also be used to publish/delete chute and register/revoke edge compute node. The edge API supports local context information of the edge compute node to chutes, such as network connection status and peripheral devices information, so that services can process and analyze the data locally. As a result, it brings the intelligence to the network of the edge.", "id": "b08f7313-052a-475e-91fc-ecf9b1117462", "level": "subsection", "origin_cites_number": 0, "parent_id": "b72e7b0d-2142-415a-8bed-1b1f88160a2d", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "ParaDrop" ], [ "subsection", "System Architecture" ] ], "subsections": [], "title": "System Architecture" }, { "cite_extract_rate": 0, "cites": [], "content": "Chute is the crucial component of the ParaDrop system, it refers to the service package that running on the ParaDrop edge compute nodes. Multiple chutes can be deployed on one edge node, and they are isolated from each other by using the container technology. Traditional cloud services can be packaged into chute to improve the user experience. For complex services containing multiple microservices, part of them can be transformed into chute while rest of the microservices can still run in the cloud. Chute also provides well supports for cloud-edge cooperation.\nChute can be seen as a Docker image containing ParaDrop and services related files. Each Chute has a configuration file that defines its requirements for various resources. ParaDrop edge compute node allocates appropriate resources according to its requirements and guarantees fairness among multi-tenants. For examples, we can set the CPU share value to 1024, that is, when there are other chutes competing for CPU resources in the system, the relative resource amount that this chute should be allocated is 1024. For maximum memory configuration, it is a hard constraint. If the data is set too low, the chute kernel may not be fully booted due to insufficient memory.\nAsides from resource information, chute configuration file also includes a runtime domain and a traffic domain. The Runtime domain describes the operations that chute can complete on its own. Take Chute.runtime information of SecCam as an example, webhosting can create a uhttpd instance based on the configured parameters and the DHCP server establishes a default version for connecting to security cameras. In many cases, devices that chute wants to interact with are probably not directly connected to the chute's network. For security concerns, ParaDrop allows developers to design and implement communication traffic rules which are wrote on the traffic domains. The rules can be implemented on the top of the firewall rules of the host network stack, allowing devices on the host LAN to connect to a specific interface to interact with chute. If users in the LAN want to obtain the data stored on chute, they can reach to the web server running on chute based on the set rules or connect through the default ParaDrop SSID in SSH mode to obtain the webpage. Compared with the traditional cloud computing mode, users can obtain data more quickly and avoid the transmission from the edge node to the cloud server.", "id": "adac1e95-6fb5-4cca-b7a1-c16f16b2589a", "level": "subsection", "origin_cites_number": 0, "parent_id": "b72e7b0d-2142-415a-8bed-1b1f88160a2d", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "ParaDrop" ], [ "subsection", "ParaDrop Chute" ] ], "subsections": [], "title": "ParaDrop Chute" }, { "cite_extract_rate": 0, "cites": [], "content": "To design a ParaDrop program, developers need to focus on two components (as shown in Figure X): 1) build tools which responsible for communicating with the rest of the ParaDrop system; 2) ParaDrop daemon instance tools (Instance tools) that responsible for running the application on various edge devices. Developers can create and manage chutes through these tools. The detailed implementation is done by the ParaDrop system and it is transparent which allowing developers to focus more on the application/service itself.\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{ParaDrop_3.png}\n\\caption{ParaDrop developer system}\\label{fig:ParaDrop_3}\n\\end{center}\n\\end{figure}\nThe basic steps to develop a chute includes:\n\\begin{enumerate}\n\\item Evaluate the applications’ functionality and performance locally;\n\\item Use the build tool to package binary files, scripts, configuration files and Docker files as chute;\n\\item Test chute at the local ParaDrop edge compute node;\n\\item Publish chute in ChuteStore and push chute to one or more edge nodes;\n\\item Use instance tool on the edge node to open the chute package and set the running environment based on the requirements.\n\\item The Docker engine creates a chute container on the edge node, including downloading the base image; installing the package; downloading the executable file and resource files;\n\\item Finally, start the developed chute.\n\\end{enumerate}\nThe ParaDrop system also offers the light chutes development method, which leverages pre-compiled base images that are optimized for different programming languages rather than building from scratch. Its compiling and installation are similar to. Although the most of the functions are similar, compared with the general chutes, the advantages of the light chutes include: higher security protection, fast installation and improved portability. Since base image is most likely already cached on the edge device, it is easy to install in a short time.\n}", "id": "e332c996-588d-441c-bafd-1d99b5403e4d", "level": "subsection", "origin_cites_number": 0, "parent_id": "b72e7b0d-2142-415a-8bed-1b1f88160a2d", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "ParaDrop" ], [ "subsection", "ParaDrop Developer Workflow" ] ], "subsections": [], "title": "ParaDrop Developer Workflow" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{sec:open-source-projects}\nBesides the designed edge computing systems for specific purposes, some open source edge computing projects have also been launched recently. The Linux Foundation published two projects, EdgeX Foundry in 2017 and Akraino Edge Statck~ in 2018. The Open Network Foundation(ONF) launched a project namely CORD (Central Office Re-architected as a Datacenter)~. The Apache Software Foundation published Apache Edgent. Microsoft published Azure IoT Edge in 2017 and announced it as an open source in 2018.\nAmong them, CORD and Akraino Edge Stack focus on providing edge cloud services; EdgeX Foundry and Apache Edgent focus on IoT and aim to solve problems which bring difficulties to practical applications of edge computing in IoT; Azure IoT Edge provides hybrid cloud-edge analytics, which helps to migrate cloud solutions to IoT devices.", "id": "e695a48d-424b-476b-aca3-58d54c720c98", "level": "section", "origin_cites_number": 2, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ] ], "subsections": [ "1446b4fa-7c0c-42d1-a70f-09fde4a71f2c", "fb60c39e-d592-4b9c-a844-9af05b80d85f", "5f68758b-531e-45bb-9d03-537df0f852c4", "62b19b9e-8f3e-4206-80e6-13afc62367cb", "af81b2b6-3749-41ab-8e6a-cce258d013d8", "e0298129-29bd-4b69-9e8c-2c1905e898d2" ], "title": "Open Source Edge Computing Projects" }, { "cite_extract_rate": 0.2, "cites": [ 8836 ], "content": "\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{CORD.pdf}\n\\caption{Hardware Architecture of CORD.}\\label{fig:CORD}\n\\end{center}\n\\end{figure} \nCORD is an open source project of ONF initiated by AT\\&T and is designed for network operators. Current network infrastructure is built with closed proprietary integrated systems provided by network equipment providers. Due to the closed property, the network capability cannot scale up and down dynamically. And the lack of flexibility results in inefficient utilization of the computing and networking resources. CORD plans to reconstruct the edge network infrastructure to build datacenters with SDN~, NFV~ and Cloud technologies. It attempts to slice the computing, storage and network resources so that these datacenters can act as clouds at the edge, providing agile services for end users.\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{CORD_SW.pdf}\n\\caption{Software Architecture of CORD.}\\label{fig:CORD_SW}\n\\end{center}\n\\end{figure}\nCORD is an integrated system built from commodity hardware and open source software. Fig.~\\ref{fig:CORD} shows the hardware architecture of CORD~. It uses commodity servers that are interconnected by a Fabric of White-box switches. White-box switch~ is a component of SDN switch, which is responsible to regulate the flow of data according to SDN controller. These commodity servers provide computing, storage resources, and the fabric of switches are used to build the network. This switching fabric is organized to a Spine-Leaf topology , a kind of flat network topology structure which adds a horizontal network structure parallel to the trunk longitudinal network structure, and then adds corresponding switching network on the horizontal structure. Comparing to traditional three-tier network topology, it can provide scalable throughput for greater East-to-West network traffic, that is, traffic coming from network diagram drawings that usually depict local area network (LAN) traffic horizontally. In addition, specialized access hardware is required to connect subscribers. The subscribers can be divided into three categories for different use cases, mobile subscribers, enterprise subscribers and residential subscribers. Each category demands different access hardware due to different access technologies. In terms of software, Fig.~\\ref{fig:CORD_SW} shows the software architecture of CORD~. Based on the servers and the fabric of switches, OpenStack provides with IaaS capability for CORD, it manages the compute, storage and networking resources as well as creating virtual machines and virtual networks. Docker is used to run services in containers for isolation. ONOS(Open Network Operating System) is a network operating system which is used to manage network components like the switching fabric and provide communication services to end-users. XOS provides a control plane to assemble and compose services. Other software projects provide component capabilities, for example, vRouter(Virtual Router) provides with virtual routing functionality.\nThe edge of the operator network is a sweet spot for edge computing because it connects customers with operators and is close to customers' applications as data sources. CORD takes edge computing into consideration and moves to support edge computing as a platform to provide edge cloud services (from the released version $4.1$). CORD can be deployed into three solution, M-CORD (Mobile CORD), R-CORD (Residential CORD) and E-CORD (Enterprise CORD) for different use cases. M-CORD focuses on mobile network, especially 5G network, and it plans to disaggregate and virtualize cellular network functions to enable services be created and scaled dynamically. This agility helps to provide multi-access edge services for mobile applications. For those use cases like driverless cars or drones, users can rent the edge service to run their edge applications. Similarly, R-CORD and E-CORD are designed to be agile service delivery platforms but for different users, residential and enterprise users relatively. \nSo far, the deployment of CORD is still under test among network operators, and more researches are needed to combine CORD with various edge applications.", "id": "1446b4fa-7c0c-42d1-a70f-09fde4a71f2c", "level": "subsection", "origin_cites_number": 5, "parent_id": "e695a48d-424b-476b-aca3-58d54c720c98", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "CORD" ] ], "subsections": [], "title": "CORD" }, { "cite_extract_rate": 0, "cites": [], "content": "Akraino Edge Stack, initiated by AT\\&T and now hosted by Linux Foundation, is a project to develop a holistic solution for edge infrastructure so as to support high-availability edge cloud services~. An open source software stack, as the software part of this solution, is developed for network carrier to facilitate optimal networking and workload orchestration for underlying infrastructure in order to meet the need of edge computing such as low latency, high performance, high availability, scalability and so on.\nTo provide a holistic solution, Akraino Edge Stack has a wide scope from infrastructure layer to application layer. Fig.~\\ref{fig:Akraino}~ shows the scope with three layers. In the application layer, Akraino Edge Stack wants to create a virutal network function (VNF) ecosystem and calls for edge applications. The second layer consists of middleware which supports applications in the top layer. In this layer, Akraino Edge Stack plans to develop Edge API and framework for interoperability with 3rd party Edge projects such as EdgeX Foundry. At the bottom layer, Akraino Edge Stack intends to develop an open source software stack for the edge infrastructure in collaboration with upstream communities. It interfaces with and maximizes the use of existing open source projects such as Kubernetes, OpenStack and so on. Akraino Edge Stack provides different edge use cases with blueprints, which are declarative configurations of entire stack including hardware, software, point of delivery, etc~. The application domains of these blueprints start from Telco industry and are expected to applied in more domains like Enterprise and industrial IoT. Now Akraino Edge Stack has put forward several blueprints such as Micro-MEC and Edge Media Processing. Micro-MEC intends to develop a new service infrastructure for smart cities, which enables developing services for smart city and has high data capacity for citizens. Edge Media Processing intends to develop a network cloud to enable real-time media processing and edge media AI analytics with low latency.\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{Akraino.png}\n\\caption{Akraino Edge Stack's Scope.}\\label{fig:Akraino}\n\\end{center}\n\\end{figure}\nAs an emerging project, Akraino Edge Stack has been taken to execution since August 2018. Thus more researches need to be done with the development of this project.", "id": "fb60c39e-d592-4b9c-a844-9af05b80d85f", "level": "subsection", "origin_cites_number": 1, "parent_id": "e695a48d-424b-476b-aca3-58d54c720c98", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Akraino Edge Stack" ] ], "subsections": [], "title": "Akraino Edge Stack" }, { "cite_extract_rate": 0.5, "cites": [ 8836 ], "content": "EdgeX Foundry is a standardized interoperability framework for IoT edge computing, whose sweet spots are edge nodes such as gateways, hubs, and routers~. It can connect with various sensors and devices via different protocols, manage them and collect data from them, and export the data to a local application at the edge or the cloud for further processing. EdgeX is designed to be agnostic to hardware, CPU, operating system, and application environment. It can run natively or run in docker containers.\nFig.~\\ref{fig:EdgeX}~ shows the architecture of EdgeX Foundry. ``South side'' at the bottom of the figure includes all IoT objects, and the edge of the network that communicates directly with those devices, sensors, actuators and other IoT objects to collect the data from them. Relatively, ``north side'' at the top of the figure includes the Cloud (or Enterprise system) where data are collected, stored, aggregated, analyzed, and turned into information, and the part of the network that communicates with the Cloud. EdgeX Foundry connects these two sides regardless of the differences of hardware, software and network. EdgeX tries to unify the manipulation method of the IoT objects from the south side to a common API, so that those objects can be manipulated in the same way by the applications of north side.\nEdgeX uses a Device Profile to describe a south side object. A Device Profile defines the type of the object, the format of data that the object provides, the format of data to be stored in EdgeX and the commands used to manipulate this object. Each Device Profile involves a Device Service, which is a service that converts the format of the data, and translates the commands into instructions that IoT objects know how to execute. EdgeX provides SDK for developers to create Device Services, so that it can support for any combination of device interfaces and protocols by programming.\n\\begin{figure*}[!tb]\n\\begin{center}\n\\includegraphics[width=0.65\\textwidth]{EdgeX2.png}\n\\caption{Architecture of EdgeX Foundry.}\\label{fig:EdgeX}\n\\end{center}\n\\end{figure*}\nEdgeX consists of a collection of microservices, which allows services to scale up and down based on device capability. These microservices can be grouped into four service layers and two underlying augmenting system services, as depicted in Fig.~\\ref{fig:EdgeX}. The four service layers include Device Services Layer, Core Services Layer, Supporting Services Layer and Export Services Layer, respectively; the two underlying augmenting system services are System Management and Security, respectively. Each of the six layers consists of several components and all components use a common Restful API for configuration.\n\\textit{1) Device Services Layer:}\nThis layer consists of Device Services. According to the Device Profiles, Device Service Layer converts the format of the data, sends them to Core Services Layer, and translates the command requests from the Core Services Layer.\n\\textit{2) Core Services Layer:}\nThis layer consists of four components: Core Data, Command, Metadata, and Registry \\& Configuration. Core Data is a persistence repository as well as a management service. It stores and manages the data collected from the south side objects. Command is a service to offer the API for command requests from the north side to Device Services. Metadata is a repository and management service for metadata about IoT objects. For example, the Device Profiles are uploaded and stored in Metadata. Registry \\& Configuration provides centralized management of configuration and operating parameters for other microservices. \n\\textit{3) Supporting Services Layer:}\nThis layer is designed to provide edge analytics and intelligence~. Now the Rules Engine, Alerting and Notification, Scheduling and Logging microservices are implemented. A target range of data can be set to trigger a specific device actuation as a rule and Rules Engine helps realize the rule by monitoring the incoming data. Alerting and Notifications can send notifications or alerts to another system or person by email, REST callback or other methods when an urgent actuation or a service malfunction happens. The scheduling module can set up a timer to regularly clean up the stale data. Logging is used to record the running information of EdgeX.\n\\textit{4) Export Services Layer:}\nThis layer connects EdgeX with North Side and consists of Client Registration and Export Distribution. Client Registration enables clients like a specific cloud or a local application to register as recipients of data from Core Data. Export Distribution distributes the data to the Clients registered in Client Registration. \n\\textit{5) System Management and Security:}\nSystem Management provides management operations including installation, upgrade, starting, stopping and monitoring, as EdgeX is scalable and can be deployed dynamically. Security is designed to protect the data and command of IoT objects connected with EdgeX Foundry.\nEdgeX is designed for the user cases dealing with multitudes of sensors or devices, such as automated factories, machinery systems and lots of other cases in IoT. Now EdgeX Foundry is in the rapid upgrading phase, and more features will be added in future releases. An EdgeX UI is in development as a web-based interface to add and manage the devices.", "id": "5f68758b-531e-45bb-9d03-537df0f852c4", "level": "subsection", "origin_cites_number": 2, "parent_id": "e695a48d-424b-476b-aca3-58d54c720c98", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "EdgeX Foundry" ] ], "subsections": [], "title": "EdgeX Foundry" }, { "cite_extract_rate": 0, "cites": [], "content": "Apache Edgent, which was known as Apache Quarks previously, is an Apache Incubator project at present. It is an open source programming model for lightweight runtime data analytics, used in small devices such as routers and gateways at the edge. Apache Edgent focuses on data analytics at the edge, aiming to accelerate the development of data analysis. \n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{Edgent.pdf}\n\\caption{Model of the Edgent Applications.}\\label{fig:Edgent}\n\\end{center}\n\\end{figure}\nAs a programming model, Edgent provides API to build edge applications. Fig.~\\ref{fig:Edgent} illustrates the model of the Edgent applications. Edgent uses a topology as a graph to represent the processing transformation of streams of data which are abstracted to a Tstream class. A connector is used to get streams of data from external entities such as sensors and devices in physical world, or to send streams of data to back-end systems like a cloud. The primary API of Edgent is responsible for data analysis. The streams of data can be filtered, split, transformed or processed by other operations in a topology. Edgent uses a provider to act as a factory to create and execute topologies. To build an Edgent application, users should firstly get a provider, then create a topology and add the processing flow to deal with the streams of data, and finally submit the topology. The deployment environments of Edgent are Java 8, Java 7 and Android. \nEdgent provides APIs for sending data to back-end systems and now supports MQTT, IBM Watson IoT Platform, Apache Kafka and custom message hubs. Edgent applications analyze the data from sensors and devices, and send the essential data to the back-end system for further analysis. For IoT use cases, Edgent helps reduce the cost of transmitting data and provide local feedback. \nEdgent is suitable for use cases in IoT such as intelligent transportation, automated factories and so on. In addition, the data in Edgent applications are not limited to sensor readings, they can also be files or logs. Therefore, Edgent can be applied to other use cases. For example, it can perform local data analysis when embedded in application servers, where it can analyze error logs without impacting network traffic~.", "id": "62b19b9e-8f3e-4206-80e6-13afc62367cb", "level": "subsection", "origin_cites_number": 1, "parent_id": "e695a48d-424b-476b-aca3-58d54c720c98", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Apache Edgent" ] ], "subsections": [], "title": "Apache Edgent" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{subsec:AzureIoT}\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.5\\textwidth]{AzureIoT.pdf}\n\\caption{Diagram of Azure IoT Edge.}\\label{fig:AzureIoT}\n\\end{center}\n\\end{figure}\nAzure IoT Edge, provided by Microsoft Azure as a cloud service provider, tries to move cloud analytics to edge devices. These edge devices can be routers, gateways or other devices which can provide computing resources. The programming model of Azure IoT Edge is the same as that of other Azure IoT services~ in the cloud, which enables user to move their existing applications from Azure to the edge devices for lower latency. The convenience simplifies the development of edge applications. In addition, Azure services like Azure Functions, Azure Machine Learning and Azure Stream Analytics can be used to deploy complex tasks on the edge devices such as machine learning, image recognition and other tasks about artificial intelligence.\nAzure IoT Edge consists of three components: IoT Edge modules, IoT Edge runtime and a cloud-based interface as depicted in Fig.~\\ref{fig:AzureIoT}. The first two components run on edge devices, and the last one is an interface in the cloud. IoT Edge modules are containerized instances running the customer code or Azure services. IoT Edge runtime manages these modules. The cloud-based interface is used to monitor and manage the former two components, in other words, monitor and manage the edge devices.\nIoT Edge modules are the places that run specific applications as the units of execution. A module image is a docker image containing the user code. A module instance, as a docker container, is a unit of computation running the module image. If the resources at the edge devices are sufficient, these modules can run the same Azure services or custom applications as in the cloud because of the same programming model. In addition, these modules can be deployed dynamically as Azure IoT Edge is scalable.\nIoT Edge runtime acts as a manager on the edge devices. It consists of two modules: IoT Edge hub and IoT Edge agent. IoT Edge hub acts as a local proxy for IoT Hub which is a managed service and a central message hub in the cloud. As a message broker, IoT Edge hub helps modules communicate with each other, and transport data to IoT Hub. IoT Edge agent is used to deploy and monitor the IoT Edge modules. It receives the deployment information about modules from IoT Hub, instantiates these modules, and ensures they are running, for example, restarts the crashed modules. In addition, it reports the status of the modules to the IoT hub.\nIoT Edge cloud interface is provided for device management. By this interface, users can create edge applications, then send these applications to the device and monitor the running status of the device. This monitoring function is useful for use cases with massive devices, where users can deploy applications to devices on a large scale and monitor these devices.\nA simple deployment procedure for applications is that: users choose a Azure service or write their own code as an application, build it as an IoT Edge module image, and deploy this module image to the edge device with the help of the IoT Edge interface. Then the IoT Edge receives the deployment information, pulls the module image, and instantiates the module instance.\nAzure IoT Edge has wide application areas. Now it has application cases on intelligent manufacturing, irrigation system, drone management system and so on. It is worth noting that Azure IoT Edge is open-source but the Azure services like Azure Functions, Azure Machine Learning and Azure Stream are charged.", "id": "af81b2b6-3749-41ab-8e6a-cce258d013d8", "level": "subsection", "origin_cites_number": 1, "parent_id": "e695a48d-424b-476b-aca3-58d54c720c98", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Azure IoT Edge" ] ], "subsections": [], "title": "Azure IoT Edge" }, { "cite_extract_rate": 0, "cites": [], "content": "We summarize the features of the above open source edge computing systems in Table~\\ref{tab:characterEdge}. Then we compare them from different aspects in Table~\\ref{tab:characterEdge}, including the main purpose of the systems,application area, deployment, target user, virtualization technology, system characteristic, limitations, scalability and mobility. We believe such comparisons give better understandings of current open source edge computing systems. \n\\begin{table*}[htb]\n\\centering\n\\caption{Comparison of Open Edge System Characteristics}\n\\label{tab:characterEdge}\n\\begin{tabular}{cccccc}\n\\hline\n\\textbf{Aspect} & \\textbf{EdgeX Foundry} & \\begin{tabular}[c]{@{}c@{}}\\textbf{Azure IoT}\\\\ \\textbf{Edge}\\end{tabular} & \\textbf{Apache Edgent} & \\textbf{CORD} & \\begin{tabular}[c]{@{}c@{}}\\textbf{Akraino}\\\\ \\textbf{Edge Stack}\\end{tabular}\n \\\\ \\hline\n\\begin{tabular}[c]{@{}c@{}}User access\\\\ interface\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Restful API\\\\ or EdgeX UI\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Web service,\\\\ Command-line\\end{tabular} & API & \\begin{tabular}[c]{@{}c@{}}API or\\\\ XOS-GUI\\end{tabular} & N/A \\\\ \\hline\nOS support & Various OS & Various OS & Various OS & Ubuntu & Linux \\\\ \\hline\n\\begin{tabular}[c]{@{}c@{}}Programming\\\\ framework\\end{tabular} & Not provided & \\begin{tabular}[c]{@{}c@{}}Java, .NET, C,\\\\ Python, etc.\\end{tabular} & Java & \\begin{tabular}[c]{@{}c@{}}Shell script,\\\\ Python\\end{tabular} &N/A\n\\\\ \\hline\nMain purpose & \\begin{tabular}[c]{@{}c@{}}Provide with\\\\ Interoperability\\\\ for IoT edge\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Support hybrid\\\\ cloud-edge\\\\ analytics\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Accelerate the \\\\ development\\\\ process of\\\\ data analysis\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Transform edge of\\\\ the operator network\\\\ into agile service\\\\ delivery platforms\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Support edge\\\\ clouds with an\\\\ open source\\\\ software stack\\end{tabular} \\\\ \\hline\nApplication area & IoT & Unrestricted & IoT & Unrestricted & Unrestricted \\\\ \\hline\nDeployment & Dynamic & Dynamic & Static & Dynamic & Dynamic \\\\ \\hline\nTarget user & General users & General users & General users & Network operators & \\begin{tabular}[c]{@{}c@{}}Network\\\\ operators\\end{tabular} \\\\ \\hline\n\\begin{tabular}[c]{@{}c@{}}Virtualization \\\\ technology\\end{tabular} & Container & Container & JVM & \\begin{tabular}[c]{@{}c@{}}Virtual Machine\\\\ and Container\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Virtual Machine\\\\ and Container\\end{tabular} \\\\ \\hline\n\\begin{tabular}[c]{@{}c@{}}System\\\\ characteristics\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}A common API for\\\\ device management\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Powerful\\\\ Azure services\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}APIs for\\\\ data analytics\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Widespread\\\\ edge clouds\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Widespread\\\\ edge clouds\\end{tabular} \\\\ \\hline\nLimitation & \\begin{tabular}[c]{@{}c@{}}Lack of\\\\ programable\\\\ interface\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Azure Services\\\\ is chargeable\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Limited to\\\\ data analytics\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Unable to\\\\ be offline\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Unable to\\\\ be offline\\end{tabular} \n \\\\ \\hline\n\\begin{tabular}[c]{@{}c@{}}Scalability\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Scalable\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Scalable\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Not scalable\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Scalable\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Scalable\\end{tabular}\n \\\\ \\hline\n\\begin{tabular}[c]{@{}c@{}}Mobility\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Not support\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Not support\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Not support\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Support\\end{tabular} & \\begin{tabular}[c]{@{}c@{}}Support\\end{tabular}\n\\\\ \\hline\n\\end{tabular}\n\\end{table*}", "id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "level": "subsection", "origin_cites_number": 0, "parent_id": "e695a48d-424b-476b-aca3-58d54c720c98", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ] ], "subsections": [ "1bf4941a-3b2d-427d-a60c-eafb5647903b", "9a51feef-b020-4190-b5d7-5786064a0299", "e06a57cf-a435-43b6-82dc-5bead50804ba", "b98f232b-467f-4515-9bce-538e319df917", "6dec77a6-4f15-4759-b0bb-ba5175fd3c7a", "ff5a04f4-5ec2-4c3a-b42f-e112c6f3b1e2", "7fd26e57-2cf5-4496-a04b-c492c08696f5", "85fc4e16-f456-41a3-9214-86f76e219c7c", "365dab20-e711-47f8-a7e6-b7d0607b797d", "49fdae92-d420-4706-96a2-36269bfa7536" ], "title": "Comparative Study" }, { "cite_extract_rate": 0, "cites": [], "content": "The main purpose shows the target problem that a system tries to fix, and it is a key factor for us to choose a suitable system to run edge applications. As an interoperability framework, EdgeX Foundry aims to communicate with any sensor or device in IoT. This ability is necessary for edge applications with data from various sensors and devices. Azure IoT Edge offers an efficient solution to move the existing applications from cloud to edge, and to develop edge applications in the same way with the cloud applications. Apache Edgent helps to accelerate the development process of data analysis in IoT use cases. CORD aims to reconstruct current edge network infrastructure to build datacenters so as to provide agile network services for end-user customers. From the view of edge computing, CORD provides with multi-access edge services. Akraino Edge Stack provides an open source software stack to support high-availability edge clouds.", "id": "1bf4941a-3b2d-427d-a60c-eafb5647903b", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Main purpose" ] ], "subsections": [], "title": "Main purpose" }, { "cite_extract_rate": 0, "cites": [], "content": "EdgeX Foundry and Apache Edgent both focus on IoT edge, and EdgeX Foundry is geared toward communication with various sensors and devices, while Edgent is geared toward data analysis. They are suitable for intelligent manufacturing, intelligent transportation and smart city where various sensors and devices generate data all the time. Azure IoT Edge can be thought as the expansion of Azure Cloud. It has an extensive application area but depends on the compute resources of edge devices. Besides, it is very convenient to deploy edge applications about artificial intelligence such as machine learning and image recognition to Azure IoT Edge with the help of Azure services. CORD and Akraino Edge Stack support edge cloud services, which have no restriction on application area. If the edge devices of users don't have sufficient computing capability, these two systems are suitable for users to run resource-intensive and interactive applications in connection with operator network.", "id": "9a51feef-b020-4190-b5d7-5786064a0299", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Application area" ] ], "subsections": [], "title": "Application area" }, { "cite_extract_rate": 0, "cites": [], "content": "As for the deployment requirements, EdgeX Foundry, Apache Edgent and Azure IoT Edge are deployed in edge devices such as routers, gateways, switchers and so on. Users can deploy EdgeX Foundry by themselves, add or reduce microservices dynamically, and run their own edge applications. Differently, users need the help of cloud-based interface to deploy Azure IoT Edge and develop their edge applications. CORD and Akraino Edge Stack are designed for network operators, who need fabric switches, access devices, network cards and other related hardware apart from compute machines. Customers have no need to think about the hardware requirements and management process of the hardware, but to rent the services provided by the network operators like renting a cloud service instead of managing a physical server.", "id": "e06a57cf-a435-43b6-82dc-5bead50804ba", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Deployment" ] ], "subsections": [], "title": "Deployment" }, { "cite_extract_rate": 0, "cites": [], "content": "Though these open source systems focus on edge computing, their target users are not the same. EdgeX Foundry, Azure IoT Edge and Apache Edgent have no restriction on target users. Therefore, every developer can deploy them into local edge devices like gateways, routers and hubs. Differently, CORD and Akraino Edge Stack are created for network operators because they focus on edge infrastructure.", "id": "b98f232b-467f-4515-9bce-538e319df917", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Target user" ] ], "subsections": [], "title": "Target user" }, { "cite_extract_rate": 0, "cites": [], "content": "Virtualization technologies are widely used nowadays. Virtual machine technology can provide better management and higher utilization of resources, stability, scalability and other advantages. Container technology can provide services with isolation and agility but with negligible overhead, which can be used in edge devices~. Using OpenStack and Docker as software components, CORD and Akraino Edge Stack use both of these two technologies to support edge cloud. Different edge devices may have different hardware and software environment. For those edge systems which are deployed on edge devices, container is a good technology for services to keep independence in different environment. Therefore, EdgeX Foundry and Azure IoT Edge choose to run as docker containers. As for Edgent, Edgent applications run on JVM.", "id": "6dec77a6-4f15-4759-b0bb-ba5175fd3c7a", "level": "subsubsection", "origin_cites_number": 1, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Virtualization technology" ] ], "subsections": [], "title": "Virtualization technology" }, { "cite_extract_rate": 0, "cites": [], "content": "System characteristics show the unique features of the system, which may help users to develop, deploy or monitor their edge applications. It will save lots of workload and time if making good use of these characteristics. EdgeX Foundry provides a common API to manage the devices, and this brings great convenience to deploying and monitoring edge applications in large scale. Azure IoT Edge provides powerful Azure services to accelerate the development of edge applications. Apache Edgent provides a series of functional APIs for data analytics, which lowers the difficulty and reduces the time on developing edge analytic applications. CORD and Akraino Edge Stack provide with multi-access edge services on edge cloud. We only need to keep connection with operator network, and we can apply for these services without the need to deploy an edge computing system on edge devices by ourselves.", "id": "ff5a04f4-5ec2-4c3a-b42f-e112c6f3b1e2", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "System characteristic" ] ], "subsections": [], "title": "System characteristic" }, { "cite_extract_rate": 0, "cites": [], "content": "This subsection discusses the limitation of the latest version of them to deploy edge applications. The lastest version of EdgeX Foundry has not provided a programmable interface in its architecture for developers to write their own applications. Although EdgeX allows us to add custom implementations, it demands more workload and time. As for Azure IoT Edge, though it is open-source and free, Azure services are chargeable as commercial software. For Apache Edgent, it is lightweight and it focuses on only data analytics. As for CORD and Akraino Edge Stack, these two systems demand stable network between data sources and the operators because the edge applications are running on the edge of operator network rather than local devices.", "id": "7fd26e57-2cf5-4496-a04b-c492c08696f5", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Limitation" ] ], "subsections": [], "title": "Limitation" }, { "cite_extract_rate": 0, "cites": [], "content": "Increasing applications at edge make the network architecture more complex and the application management more difficult. Scalability is one major concern in edge computing. Among these edge computing systems, Azure IoT Edge, CORD and Akraino Edge Stack apply docker technology or virtual machine technology to support users to scale up or down their applications efficiently by adding or deleting module images. EdgeX Foundry is also a scalable platform that enables users to dynamically add or reduce microservices to adapt to the actual needs. But Apache Edgent is not scalable enough because every Edgent application is a single Java application and performance cannot be changed dynamically.", "id": "85fc4e16-f456-41a3-9214-86f76e219c7c", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Scalability" ] ], "subsections": [], "title": "Scalability" }, { "cite_extract_rate": 0, "cites": [], "content": "For EdgeX Foundry, Apache Edgent and Azure IoT Edge, once the applications are executed on some edge devices, they cannot be dynamically migrated to other devices. CORD and Akraino Edge Stack, deployed in the telecom infrastructure, support mobile edge services through mobile access network like 4G/5G. The mobility of these systems meet the need of cases such as unmanned cars and drones.", "id": "365dab20-e711-47f8-a7e6-b7d0607b797d", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Mobility" ] ], "subsections": [], "title": "Mobility" }, { "cite_extract_rate": 0, "cites": [], "content": "We discuss the choice of open-source tools from the perspective of the following three scenarios, as shown in Fig.~\\ref{fig:scenarios}.\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.48\\textwidth]{scenarios.pdf}\n\\caption{The choice of open-source tools in different scenarios.}\\label{fig:scenarios}\n\\end{center}\n\\end{figure}\nIn the first scenario, suppose IoT edge applications are running on local area network (LAN), and the local enterprise system is regarded as back-end system with no need for third-party clouds. In this case, EdgeX Foundry or Apache Edgent are favorable because they enable users to build and control their own back-end system without being bound to any specific cloud platform. Furthermore, for the sake of managing and controlling edge devices on a large scale, EdgeX Foundry is a better choice for good device management capability. If considering data analysis, Apache Edgent is preferable to EdgeX Foundry. It provides a programming model to accelerate the development of edge analytic applications and a set of powerful APIs for data processing. \nIn the second scenario, suppose the cloud services push to the edge of network to improve the quality. In this case, Auzre IoT Edge provides a convenient way for developers to migrate the applications from the cloud to the edge devices, and to leverage third-party high value services or functions when developing edge systems. Besides, AWS IoT Greengrass and Link IoT Edge, which are published by Amazon and Alibaba Cloud, are good choices as the competitive projects. More specifically, Azure IoT Edge provides Azure services such as Azure Functions, Azure Stream Analytics, and Azure Machine Learning. AWS IoT Greengrass can run AWS Lamda functions and machine learning models that are created, trained, and optimized in the cloud. Link IoT Edge provides Function Compute and other functions. Based on the application requirements, a suitable system could be chosen among them by taking account the functions they provide.\nIn the third scenario, suppose the users expect to directly leverage third-party services to deploy edge applications without any hardware or software system locally. In this case, edge systems such as CORD and Akraino Edge Stack are suitable. The users could choose one of them dependent on their application requirements. These systems are deployed by telecom operators. And telecom operators are also the providers of network services. Thus when there exist special network requirements of edge applications, these systems could satisfy the requirements. For example, edge computing applications on unmanned vehicles or drones need the support of wireless telecom networks (such as 4G/5G), in this case, a mobile edge computing service provided by these systems is a good choice.\nIn addition to the systems described above, there are other emerging open source projects. Device Management Edge, as part of Mbed IoT Platform published by ARM, is responsible for edge computing and provides the ability to access, manage and control Edge devices. KubeEdge, released by Huawei, provides edge nodes with native containerized application orchestration capabilities.", "id": "49fdae92-d420-4706-96a2-36269bfa7536", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "e0298129-29bd-4b69-9e8c-2c1905e898d2", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Open Source Edge Computing Projects" ], [ "subsection", "Comparative Study" ], [ "subsubsection", "Scenarios" ] ], "subsections": [], "title": "Scenarios" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{sec:energy-efficiency}\nIn the design of an edge computing system, energy consumption is always considered as one of the major concerns and evaluated as one of the key performance metrics. In this section, we review the energy efficiency enhancing mechanisms adopted by the state-of-the-art edge computing systems, from the view of \\emph{top cloud layer}, \\emph{middle edge server layer} and \\emph{bottom device layer}, respectively (the three-layer paradigm is illustrated by Fig.~\\ref{fig:three-layer-paradigm}).\n\\begin{figure}[!tb]\n\\begin{center}\n\\includegraphics[width=0.32\\textwidth]{three-layer-paradigm.pdf}\n\\caption{The Three-layer Edge Computing Paradigm from a Power Source View.}\\label{fig:three-layer-paradigm}\n\\end{center}\n\\end{figure}", "id": "a3a2618a-548a-4cd2-95bc-06a60c44ffb0", "level": "section", "origin_cites_number": 0, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ] ], "subsections": [ "568161be-6e40-4554-b2de-39750eac4eb0", "cef0da1f-0af4-4654-af0c-b8d0a6a63c4a", "92004c3a-1e93-4cfc-9226-2ae148fcfafd" ], "title": "Enhancing Energy Efficiency of Edge Computing Systems" }, { "cite_extract_rate": 0, "cites": [], "content": "For cloud computing, a centralized DC can comprise thousands of servers and thus consume enormous energy~. As an alternative to the cloud computing, does the edge/fog computing paradigm consume more or less energy? Different points of view have been given: some claim that decentralized data storage and processing supported by the edge computing architecture are more energy efficient~, while some others show that such distributed content delivery may consume more energy than that of the centralized way~.\nThe authors in~ give a thorough energy analysis for applications running over the centralized DC (i.e., under cloud mode) and decentralized nano DCs (i.e., under fog mode), respectively. The results indicate that the fog mode may be with a higher energy efficiency, depending on several system design factors (e.g., type of application, type of access network, and ratio of active time), and those applications that generate and distribute a large amount of data in end-user premises result in best energy saving under the fog mode.\nNote that the decentralized nano DCs or fog nodes in our context are different from the traditional CDN datacenters~. They are also designed in a decentralized way but usually with much more powerful computing/communicating/storage capacities.", "id": "568161be-6e40-4554-b2de-39750eac4eb0", "level": "subsection", "origin_cites_number": 8, "parent_id": "a3a2618a-548a-4cd2-95bc-06a60c44ffb0", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ], [ "subsection", "At the Top Cloud Layer" ] ], "subsections": [], "title": "At the Top Cloud Layer" }, { "cite_extract_rate": 0, "cites": [], "content": "At the middle layer of the edge computing paradigm, energy is also regarded as an important aspect, as the edge servers can be deployed in a domestic environment or powered by the battery (e.g., a desktop or a portable WiFi router, as shown in Fig.~\\ref{fig:three-layer-paradigm}). Thus, to provide a higher availability, many power management techniques have been applied to limit the energy consumption of edge servers, while still ensuring their performances. We give a review of two major strategies used at the edge server layer in recent edge computing systems.", "id": "cef0da1f-0af4-4654-af0c-b8d0a6a63c4a", "level": "subsection", "origin_cites_number": 0, "parent_id": "a3a2618a-548a-4cd2-95bc-06a60c44ffb0", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ], [ "subsection", "At the Middle Edge Server Layer" ] ], "subsections": [ "e29f4cf4-8bb2-43ab-9414-a3c4becd1890", "0886e58c-6c6e-466f-b652-ef84a543e8d9" ], "title": "At the Middle Edge Server Layer" }, { "cite_extract_rate": 0.25, "cites": [ 4754 ], "content": "In~, the tactical cloudlet is presented and its energy consumption when performing VM synthesis is evaluated particularly, under different cloudlet provisioning mechanisms. The results show that the largest amount of energy is consumed by i) VM synthesis due to the large payload size, and ii) on-demand VM provisioning due to the long application-ready time. Such results lead to the high energy efficiency policy: combining cached VM with cloudlet push for cloudlet provision.\nA service-oriented architecture for fog/edge computing, Fog Data, is proposed and evaluated in~. It is implemented with an embedded computer system and performs data mining and data analytics on the raw data collection from the wearable sensors (in telehealth applications). With Fog Data, orders of magnitude data are reduced for transmission, thus leading to enormous energy saving. Furthermore, Fog Data is with a low power architecture design, and even consumes much less energy than that of a Raspberry Pi.\nIn~, a performance-aware orchestrator for Docker containers, named DockerCap, is developed to meet the power consumption constraints of the edge server (fog node). Following the observe-decide-act loop structure, DockerCap is able to manage container resources at run-time and provide soft-level power capping strategies. The experiments demonstrate that the obtained results with DockerCap is comparable to that from the power capping solution provided by the hardware (Intel RAPL).\nAn energy aware edge computer architecture is designed to be portable and usable in the fieldwork scenarios in~. Based on the architecture, a high-density cluster prototype is built using the compact general-purpose commodity hardware. Power management policies are implemented in the prototype to enable the real-time energy-awareness. Through various experiments, it shows that both the load balance strategies and cluster configurations have big impacts on the system energy consumption and responsiveness.", "id": "e29f4cf4-8bb2-43ab-9414-a3c4becd1890", "level": "subsubsection", "origin_cites_number": 4, "parent_id": "cef0da1f-0af4-4654-af0c-b8d0a6a63c4a", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ], [ "subsection", "At the Middle Edge Server Layer" ], [ "subsubsection", "Low-power system design and power management" ] ], "subsections": [], "title": "Low-power system design and power management" }, { "cite_extract_rate": 0, "cites": [], "content": "Dual energy sources are employed to support the running of a fog computing based system in~, where solar power is utilized as the primary energy supply of the fog nodes. A comprehensive analytic framework is presented to minimize the long-term cost on energy consumption. Meanwhile, the framework also enables an energy-efficient data offloading (from fog nodes to the cloud) mechanism to help provide a high quality of service.\nIn~, a rack-scale green-energy powered edge infrastructure, InSURE (in-situ server system using renewable energy) is implemented for data pre-processing at the edge. InSURE can be powered by standalone (solar/wind) power and with batteries as the energy backup. Meanwhile, an energy buffering mechanism and a joint spatio-temporal power management scheme are applied, to enable efficient energy flow control from the power supply to the edge server.", "id": "0886e58c-6c6e-466f-b652-ef84a543e8d9", "level": "subsubsection", "origin_cites_number": 2, "parent_id": "cef0da1f-0af4-4654-af0c-b8d0a6a63c4a", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ], [ "subsection", "At the Middle Edge Server Layer" ], [ "subsubsection", "Green-energy powered sustainable computing" ] ], "subsections": [], "title": "Green-energy powered sustainable computing" }, { "cite_extract_rate": 1, "cites": [ 3357 ], "content": "As a well-recognized fact, the IoT devices in edge computing usually have strict energy constraints, e.g., limited battery life and energy storage. Thus, it remains a key challenge to power a great number (can up to tens of billions) of IoT devices at the edge, especially for those resource-intensive applications or services~. We review the energy saving strategies adopted at the device layer of the edge computing diagram. Specifically, we go through three major approaches to achieving high energy efficiency in different edge/fog computing systems.", "id": "92004c3a-1e93-4cfc-9226-2ae148fcfafd", "level": "subsection", "origin_cites_number": 1, "parent_id": "a3a2618a-548a-4cd2-95bc-06a60c44ffb0", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ], [ "subsection", "At the Bottom Device Layer" ] ], "subsections": [ "b5cb6b35-d6f7-494a-b718-94bbbaeaf422", "225a8b24-31ce-4561-a161-bff22095366d" ], "title": "At the Bottom Device Layer" }, { "cite_extract_rate": 0, "cites": [], "content": "As a natural idea to solve the energy poverty problem, computation offloading from the IoT devices to the edge servers or cloud has been long investigated~. It was also demonstrated that, for some particular applications or services, offloading tasks from IoT devices to more powerful ends can reduce the total energy consumption of the system, since the task execution time on powerful servers or cloud can be much shortened~. Although it increases the energy consumption of (wireless) data transmission, the tradeoff favors the offloading option as the computational demand increases~.\nHaving realized that the battery life is the primary bottleneck of handheld mobile devices, the authors in~ present MAUI, an architecture for mobile code offloading and remote execution. To reduce the energy consumption of the smartphone program, MAUI adopts a fine-grained program partitioning mechanism and minimizes the code changes required at the remote server or cloud. The ability of MAUI in energy reduction is validated by various experiments upon macro-benchmarks and micro-benchmarks. The results show that MAUI's energy saving for a resource-intensive mobile application is up to one order of magnitude, also with a significant performance improvement.\nLike MAUI~ , the authors in~ design and implement CloneCloud, a system that helps partition mobile application programs and performs strategically offloading for fast and elastic execution at the cloud end. As the major difference to MAUI, CloneCloud involves less programmer help during the whole process and only offloads particular program partitions on demand of execution, which further speeds up the program execution. Evaluation shows that CloneCloud can improve the energy efficiency of mobile applications (along with their execution efficiency) by 20 times. Similarly in~, by continuous updates of software clones in the cloud with a reasonable overhead, the offloading service can lead to energy reduction at the mobile end by a factor.\nFor computation intensive applications on resource constrained edge devices, their executions usually need to be offloaded to the cloud. To reduce the response latency of the image recognition application, Precog is presented which has been introduced in Sec.~\\ref{subsec:Cachier_and_Precog}. With the on-device recognition caches, Precog much reduces the amount of images offloading to the edge server or cloud, by predicting and prefetching the future images to be recognized.", "id": "b5cb6b35-d6f7-494a-b718-94bbbaeaf422", "level": "subsubsection", "origin_cites_number": 6, "parent_id": "92004c3a-1e93-4cfc-9226-2ae148fcfafd", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ], [ "subsection", "At the Bottom Device Layer" ], [ "subsubsection", "Computation offloading to edge servers or cloud" ] ], "subsections": [], "title": "Computation offloading to edge servers or cloud" }, { "cite_extract_rate": 0, "cites": [], "content": "For energy saving of the massive devices at the edge, besides offloading their computational tasks to more powerful ends, there is also a great potential via sophisticated collaboration and cooperation among the devices themselves. Particularly, when the remote resources from the edge server or cloud are unavailable, it is critical and non-trivial to complete the edge tasks while without violating the energy constraint.\nPCloud is presented in~ to enhance the capability of individual mobile devices at the edge. By seamless using available resources from the nearby devices, PCloud forms a ‘personal cloud’ to serve end users whenever the cloud resources are difficult to access, where device participation is guided in a privacy-preserving manner. The authors show that, by leveraging multiple nearby device resources, PCloud can much reduce the task execution time as well as energy consumption. For example, in the case study of ‘neighborhood watch with face recognition’, the results show a $74\\%$ reduction in energy consumption on a PCloud vs. on a single edge device. Similar to PCloud, the concept of mobile device cloud (MDC) is proposed in~, where computational offloading is also adopted among the mobile devices. It shows that the energy efficiency (gain) is increased by $26\\%$ via offloading in MDC. The authors of~ propose an adaptive method to dynamically discovery available nearby resource in heterogeneous networks, and perform automatic transformation between centralized and flooding strategies to save energy.\nAs current LTE standard is not optimized to support a large simultaneous access of IoT devices, the authors in~ propose an improved memory replication architecture and protocol, REPLISON, for computation offloading of the massive IoT devices at the edge. REPLISON improves the memory replication performance through an LTE-optimized protocol, where Device-to-Device (D2D) communication is applied as an important supporting technology (to pull memory replicas from IoT devices). The total energy consumption of REPLISON is generally worse than the conventional LTE scenario, as it needs more active devices. However, the energy consumed per device during a single replicate transmission is much less. With further evaluation results, it shows that REPLISOM has an energy advantage over the conventional LTE scenarios as long as the size of replica is sufficiently small.\nFor IoT devices distributed at the edge, the authors in~ leverage the software agents running on the IoT devices to establish an integrated multi-agent system (MAS). By sharing data and information among the mobile agents, edge devices are able to collaborate with each other and improve the system energy efficiency in executing distributed opportunistic applications. Upon the experimental platform with 100 sensor nodes and 20 smartphones as edge devices, the authors show the great potential of data transmission reduction with MAS. This leads to a significant energy saving, from $15\\%$ to $66\\%$, under different edge computing scenarios. As another work applying data reduction for energy saving, CAROMM~ employs a change detect technique (LWC algorithm) to control the data transmission of IoT devices, while maintaining the data accuracy.", "id": "225a8b24-31ce-4561-a161-bff22095366d", "level": "subsubsection", "origin_cites_number": 6, "parent_id": "92004c3a-1e93-4cfc-9226-2ae148fcfafd", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Enhancing Energy Efficiency of Edge Computing Systems" ], [ "subsection", "At the Bottom Device Layer" ], [ "subsubsection", "Collaborated devices control and resource management" ] ], "subsections": [], "title": "Collaborated devices control and resource management" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{sec:deep-learning-opt}\nIn the past decades, we have witnessed the burgeoning of machine learning, especially deep learning based applications which have changed human being's life. With complex structures of hierarchical layers to capture features from raw data, deep learning models have shown outstanding performances in those novel applications, such as machine translation, object detection, and smart question and answer systems.\nTraditionally, most deep learning based applications are deployed on a remote cloud center, and many systems and tools are designed to run deep learning models efficiently on the cloud. Recently, with the rapid development of edge computing, the deep learning functions are being offloaded to the edge. Thus it calls for new techniques to support the deep learning models at the edge. This section classifies these technologies into three categories: systems and toolkits, deep learning packages, and hardware.", "id": "478327fb-c0c5-49d5-ae91-7fa642949abc", "level": "section", "origin_cites_number": 0, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Deep Learning Optimization at the Edge" ] ], "subsections": [ "98b9a8e8-5e84-4189-aa8d-48851fdccc02", "083ff676-0bf0-44e0-859d-63889cdc7ccb", "e71e3ef2-27ab-4b8d-8ebe-b33c2376117c" ], "title": "Deep Learning Optimization at the Edge" }, { "cite_extract_rate": 0.33333333333333304, "cites": [ 7842 ], "content": "Building systems to support deep learning at the edge is currently a hot topic for both industry and academy. There are several challenges when offloading state-of-the-art AI techniques on the edge directly, including computing power limitation, data sharing and collaborating, and mismatch between edge platform and AI algorithms. To address these challenges, OpenEI is proposed as an Open Framework for Edge Intelligence~. OpenEI is a lightweight software platform to equip edges with intelligent processing and data sharing capability. OpenEI consists of three components: a package manager to execute the real-time deep learning task and train the model locally, a model selector to select the most suitable model for different edge hardware, and a library including a RESTFul API for data sharing. The goal of OpenEI is that any edge hardware will has the intelligent capability after deploying it. \nIn the industry, some top-leading tech-giants have published several projects to move the deep learning functions from the cloud to the edge. Except Microsoft published Azure IoT Edge which have been introduced in Sec.~\\ref{subsec:AzureIoT}, Amazon and Google also build their services to support deep learning on the edge.\nTable~\\ref{table: Comparison of Deep learning Systems on Edge} summarizes the features of the systems which will be discussed below.\nAmazon Web Services (AWS) has published IoT Greengrass ML Inference~ after IoT Greengrass. AWS IoT Greengrass ML Inference is a software to support machine learning inferences on local devices. With AWS IoT Greengrass ML Inference, connected IoT devices can run AWS Lambda functions and have the flexibility to execute predictions based on those deep learning models created, trained, and optimized in the cloud. AWS IoT Greengrass consists of three software distributions: AWS IoT Greengrass Core, AWS IoT Device SDK, and AWS IoT Greengrass SDK. Greengrass is flexible for users as it includes a pre-built TensorFlow, Apache MXNet, and Chainer package, and it can also work with Caffe2 and Microsoft Cognitive Toolkit. \nCloud IoT Edge~ extends Google Cloud's data processing and machine learning to edge devices by taking advantages of Google AI products, such TensorFlow Lite and Edge TPU. Cloud IoT Edge can either run on Android or Linux-based operating systems. It is made up of three components: Edge Connect ensures the connection to the cloud and the updates of software and firmware, Edge ML runs ML inference by TensorFlow Lite, and Edge TPU specific designed to run TensorFlow Lite ML models. Cloud IoT Edge can satisfy the real-time requirement for the mission-critical IoT applications, as it can take advantages of Google AI products (such as TensorFlow Lite and Edge TPU) and optimize the performance collaboratively.\n\\begin{table*}[!htp]\n\t\\caption{Comparison of Deep learning Systems on Edge}\n\t\\label{table: Comparison of Deep learning Systems on Edge}\n\t\\begin{tabular}{ l l l l }\n\t\t\\hline\n\t\t\\textbf{Features} & \\textbf{AWS IoT Greengrass} & \\textbf{Azure IoT Edge} & \\textbf{Cloud IoT Edge} \\\\ \\hline\n\t\tDeveloper & Amazon & Microsoft & Google \\\\ \\hline\n\t\tComponents & \\begin{tabular}[c]{@{}l@{}}IoT Greengrass Core, IoT Device SDK, \\\\ IoT Greengrass SDK\\end{tabular} & \\begin{tabular}[c]{@{}l@{}}IoT Edge modules, IoT Edge runtime, \\\\ Cloud-based interface\\end{tabular} & \\begin{tabular}[c]{@{}l@{}}Edge Connect, Edge ML, Edge TPU\\end{tabular} \\\\ \\hline\n\t\tOS & Linux, macOS, Windows & Windows, Linux, macOS & Linux, macOS, Windows, Android \\\\ \\hline\n\t\tTarget device & Multiple platforms (GPU-based, Raspberry Pi) & Multiple platforms & TPU\\\\ \\hline\n\t\tCharacteristic & Flexible & Windows friendly & Real-time \\\\ \\hline\n\t\\end{tabular}\n\\end{table*}", "id": "98b9a8e8-5e84-4189-aa8d-48851fdccc02", "level": "subsection", "origin_cites_number": 3, "parent_id": "478327fb-c0c5-49d5-ae91-7fa642949abc", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Deep Learning Optimization at the Edge" ], [ "subsection", "Systems and Toolkits" ] ], "subsections": [], "title": "Systems and Toolkits" }, { "cite_extract_rate": 0.45454545454545403, "cites": [ 3416, 2672, 2671, 7199, 4755 ], "content": "Many deep learning packages have been widely used to deliver the deep learning algorithms and deployed on the cloud data centers, including TensorFlow~, Caffe~, PyTorch~, and MXNet~. Due to the limitations of computing resources at the edge, the packages designed for the cloud are not suitable for edge devices. Thus, to support data processing with deep learning models at the edge, several edge-based deep learning frameworks and tools have been released. In this section, we introduce TensorFlow Lite, Caffe2, PyTorch, MXNet, CoreML~, and TensorRT~, whose features are summarized in Tables \\ref{table: Comparison of Deep Learning Frameworks on Edge}.\nTensorFlow Lite~ is TensorFlow's lightweight solution which is designed for mobile and edge devices. TensorFlow is developed by Google in 2016 and becomes one of the most widely used deep learning frameworks in cloud data centers. To enable low-latency inference of on-device deep learning models, TensorFlow Lite leverages many optimization techniques, including optimizing the kernels for mobile apps, pre-fused activations, and quantized kernels that allow smaller and faster (fixed-point math) models. \nFacebook published Caffe2~ as a lightweight, modular, and scalable framework for deep learning in 2017. Caffe2 is a new version of Caffe which is first developed by UC Berkeley AI Research (BAIR) and community contributors. Caffe2 provides an easy and straightforward way to play with the deep learning and leverage community contributions of new models and algorithms. Comparing with the original Caffe framework, Caffe2 merges many new computation patterns, including distributed computation, mobile, reduced precision computation, and more non-vision use cases. Caffe2 supports multiple platforms which enable developers to use the power of GPUs in the cloud or at the edge with cross-platform libraries.\nPyTorch~ is published by Facebook. It is a python package that provides two high-level features: tensor computation with strong GPU acceleration and deep Neural Networks built on a tape-based auto-grad system. Maintained by the same company (Facebook), PyTorch and Caffe2 have their own advantages. PyTorch is geared toward research, experimentation and trying out exotic neural networks, while caffe2 supports more industrial-strength applications with a heavy focus on the mobile. In 2018, Caffe2 and PyTorch projects merged into a new one named PyTorch 1.0, which would combine the user experience of the PyTorch frontend with scaling, deployment and embedding capabilities of the Caffe2 backend.\nMXNet~ is a flexible and efficient library for deep learning. It was initially developed by the University of Washington and Carnegie Mellon University, to support CNN and long short-term memory networks (LSTM). In 2017, Amazon announced MXNet as its choice of deep learning framework. MXNet places a special emphasis on speeding up the development and deployment of large-scale deep neural networks. It is designed to support multiple different platforms (either cloud platforms or the edge ones) and can execute training and inference tasks. Furthermore, other than the Windows, Linux, and OSX operating systems based devices, it also supports the Ubuntu Arch64 and Raspbian ARM based operating systems.\nCoreML~ is a deep learning framework optimized for on-device performance at memory footprint and power consumption. Published by Apple, users can integrate the trained machine learning model into Apple products, such as Siri, Camera, and QuickType. CoreML supports not only deep learning models, but also some standard models such as tree ensembles, SVMs, and generalized linear models. Built on top of low level technologies, CoreML aims to make full use of the CPU and GPU capability and ensure the performance and efficiency of data processing.\nThe platform of TensorRT~ acts as a deep learning inference to run the models trained by TensorFlow, Caffe, and other frameworks. Developed by NVIDIA company, it is designed to reduce the latency and increase the throughput when executing the inference task on NVIDIA GPU. To achieve computing acceleration, TensorRT leverages several techniques, including weight and activation precision calibration, layer and tensor fusion, kernel auto-tuning, dynamic tensor memory, and multi-stream execution.\nConsidering the different performance of the packages and the diversity of the edge hardware, it is challenging to choose a suitable package to build edge computing systems. To evaluate the deep learning frameworks at the edge and provide a reference to select appropriate combinations of package and edge hardware, pCAMP~ is proposed. It compares the packages' performances (w.r.t. the latency, memory footprint, and energy) resulting from five edge devices and observes that no framework could win over all the others at all aspects.\nIt indicates that there is much room to improve the frameworks at the edge. Currently, developing a lightweight, efficient and high-scalability framework to support diverse deep learning modes at the edge cannot be more important and urgent.\nIn addition to these single-device based frameworks, more researchers focus on distributed deep learning models over the cloud and edge. DDNN~ is a distributed deep neural network architecture across cloud, edge, and edge devices. DDNN maps the sections of a deep neural network onto different computing devices, to minimize communication and resource usage for devices and maximize usefulness of features extracted from the cloud. \nNeurosurgeon~ is a lightweight scheduler which can automatically partition DNN computation between mobile devices and datacenters at the granularity of neural network layers. By effectively leveraging the resources in the cloud and at the edge, Neurosurgeon achieves low computing latency, low energy consumption, and high traffic throughput. \n\\begin{table*}[!htp]\n\t\\centering\n\t\\caption{Comparison of Deep Learning Packages on Edge}\n\t\\label{table: Comparison of Deep Learning Frameworks on Edge}\n\t\\begin{tabular}{ l l l l l l l }\n\t\t\\hline\n\t\t\\textbf{Features} & \\textbf{TensorFlow Lite} & \\textbf{Caffe2} & \\textbf{PyTorch} & \\textbf{MXNet} & \\textbf{CoreML} & \\textbf{TensorRT} \\\\ \\hline\n\t\tDeveloper & Google & Facebook & Facebook & DMLC, Amazon & Apple & NVIDIA \\\\ \\hline\n\t\t\\begin{tabular}[c]{@{}l@{}}Open Source \\\\ License\\end{tabular} & Apache-2.0 & Apache-2.0 & BSD & Apache-2.0 & Not open source & Not open source \\\\ \\hline\n\t\tTask & Inference & Training, Inference & Training, Inference & Training, Inference & Inference & Inference \\\\ \\hline\n\t\tTarget Device & \\begin{tabular}[c]{@{}l@{}}Mobile and \\\\ embedded device\\end{tabular} & Multiple platform & Multiple platform & Multiple platform & Apple devices & NVIDIA GPU \\\\ \\hline\n\t\tCharacteristic & Latency & \\begin{tabular}[c]{@{}l@{}}Lightweight, modular, \\\\ and scalable\\end{tabular} & Research & \\begin{tabular}[c]{@{}l@{}}Large-scale \\\\ deep neural networks\\end{tabular} & \\begin{tabular}[c]{@{}l@{}}Memory footprint and \\\\ power consumption.\\end{tabular} & \\begin{tabular}[c]{@{}l@{}}Latency and\\\\ throughput\\end{tabular} \\\\ \\hline\n\t\\end{tabular}\n\\end{table*}", "id": "083ff676-0bf0-44e0-859d-63889cdc7ccb", "level": "subsection", "origin_cites_number": 11, "parent_id": "478327fb-c0c5-49d5-ae91-7fa642949abc", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Deep Learning Optimization at the Edge" ], [ "subsection", "Deep Learning Packages" ] ], "subsections": [], "title": "Deep Learning Packages" }, { "cite_extract_rate": 0, "cites": [], "content": "The hardware designed specifically for deep learning can strongly support edge computing. Thus, we further review relevant hardware systems and classify them into three categories: FPGA-based hardware, GPU-based hardware, and ASIC.", "id": "e71e3ef2-27ab-4b8d-8ebe-b33c2376117c", "level": "subsection", "origin_cites_number": 0, "parent_id": "478327fb-c0c5-49d5-ae91-7fa642949abc", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Deep Learning Optimization at the Edge" ], [ "subsection", "Hardware System" ] ], "subsections": [ "bb0c2551-78bb-4b6f-b8fd-cff2ffbb6a59", "8cf9bef2-b337-4bcf-b88f-1c2fb48cab5e", "5cde47d9-ebf3-4868-b0ed-359a6b64f5ef" ], "title": "Hardware System" }, { "cite_extract_rate": 0.33333333333333304, "cites": [ 4756 ], "content": "A field-programmable gate array (FPGA) is an integrated circuit and can be configured by the customer or designer after manufacturing. FPGA based accelerators can achieve high performance computing with low energy, high parallelism, high flexibility, and high security~.\n implements a CNN accelerator on a VC707 FPGA board. The accelerator focuses on solving the problem that the computation throughput does not match the memory bandwidth well. By quantitatively analyzing the two factors using various optimization techniques, the authors provide a solution with better performance and lower FPGA resource requirement, and their solution achieves a peak performance of $61.62$ GOPS under a $100MHz$ working frequency. \nFollowing the above work, Qiu et al.~ propose a CNN accelerator designed upon the embedded FPGA, Xilinx Zynq ZC706, for large-scale image classification. It presents an in-depth analysis of state-of-the-art CNN models and shows that Convolutional layers are computational-centric and Fully-Connected layers are memory-centric. The average performances of the CNN accelerator at convolutional layers and the full CNN are $187.8$ GOPS and $137.0$ GOPS under a $150MHz$ working frequency, respectively, which outperform previous approaches significantly.\nAn efficient speech recognition engine (ESE) is designed to speed up the predictions and save energy when applying the deep learning model of LSTM. ESE is implemented in a Xilinx XCKU060 FPGA opearting at $200MHz$. For the sparse LSTM network, it can achieve 282GOPS, corresponding to a $2.52$ TOPS on the dense LSTM network. Besides, energy efficiency improvements of $40x$ and $11.5x$ are achieved, respectively, compared with the CPU and GPU based solution.", "id": "bb0c2551-78bb-4b6f-b8fd-cff2ffbb6a59", "level": "subsubsection", "origin_cites_number": 3, "parent_id": "e71e3ef2-27ab-4b8d-8ebe-b33c2376117c", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Deep Learning Optimization at the Edge" ], [ "subsection", "Hardware System" ], [ "subsubsection", "FPGA-based Hardware" ] ], "subsections": [], "title": "FPGA-based Hardware" }, { "cite_extract_rate": 0, "cites": [], "content": "GPU can execute parallel programs at a much higher speed than CPU, which makes it fit for the computational paradigm of deep learning algorithms. Thus, to run deep learning models at the edge, building the hardware platform with GPU is a must choice. Specifically, NVIDIA Jetson TX2 and DRIVE PX2 are two representative GPU-based hardware platforms for deep learning.\nNVIDIA Jetson TX2~ is an embedded AI computing device which is designed to achieve low latency and high power-efficient. It is built upon an NVIDIA Pascal GPU with 256 CUDA cores, an HMP Dual Denver CPU and a Qualcomm ARM CPU. It is loaded with 8GB of memory and 59.7GB/s of memory bandwidth and the power is about 7.5 watts. The GPU is used to execute the deep learning task and CPUs are used to maintain general tasks. It also supports the NVIDIA Jetpack SDK which includes libraries for deep learning, computer vision, GPU computing, and multimedia processing.\nNVIDIA DRIVE PX~ is designed as the AI supercomputer for autonomous driving. The architecture is available in a variety of configurations, from the mobile processor operating at 10 watts to a multi-chip AI processors delivering 320 TOPS. It can fuse data from multiple cameras, as well as lidar, radar, and ultrasonic sensors.", "id": "8cf9bef2-b337-4bcf-b88f-1c2fb48cab5e", "level": "subsubsection", "origin_cites_number": 2, "parent_id": "e71e3ef2-27ab-4b8d-8ebe-b33c2376117c", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Deep Learning Optimization at the Edge" ], [ "subsection", "Hardware System" ], [ "subsubsection", "GPU-based hardware" ] ], "subsections": [], "title": "GPU-based hardware" }, { "cite_extract_rate": 0, "cites": [], "content": "Application-Specific Integrated Circuit (ASIC) is the integrated circuit which supports customized design for a particular application, rather than the general-purpose use. ASIC is suitable for the edge scenario as it usually has a smaller size, lower power consumption, higher performance, and higher security than many other circuits. Researchers and developers design ASIC to meet the computing pattern of deep learning.\nDianNao family~ is a series of hardware accelerators designed for deep learning models, including DianNao, DaDianNao, ShiDianNao, and PuDianNao. They investigate the accelerator architecture to minimize memory transfers as efficiently as possible. Different from other accelerators in DianNao family, ShiDianNao~ focuses on image applications in embedded systems which are widely used in edge computing scenarios. For the CNN-based deep learning models, it provides a computing speed $30x$ faster than that of NVIDIA K20M GPU, with an body area of $4.86 mm^2$ and a power of $320 mW$.\nEdge TPU~ is Google's purpose-built ASIC for edge computing. It augments Google's Cloud TPU and Cloud IoT to provide an end-to-end infrastructure and facilitates the deployment of customers' AI-based solutions. In addition, Edge TPU can combine the custom hardware, open software, and state-of-the-art AI algorithms to achieve high performance with a small physical area and low power consumption.", "id": "5cde47d9-ebf3-4868-b0ed-359a6b64f5ef", "level": "subsubsection", "origin_cites_number": 3, "parent_id": "e71e3ef2-27ab-4b8d-8ebe-b33c2376117c", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Deep Learning Optimization at the Edge" ], [ "subsection", "Hardware System" ], [ "subsubsection", "ASIC" ] ], "subsections": [], "title": "ASIC" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{sec:key-design-issues}\nThe edge computing system manages various resources along the path from the cloud center to end devices, shielding the complexity and diversity of hardware and helping developers quickly design and deploy novel applications. To fully leverages the advantages, we discuss the following key issues which need to be paid attention when analyzing and designing a new edge computing system.", "id": "aa0b39df-82ae-4dab-a3c4-17e99ed6d2be", "level": "section", "origin_cites_number": 0, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ] ], "subsections": [ "efe4a2f1-018b-4922-b876-e75becde3cd5" ], "title": "Key Design Issues" }, { "cite_extract_rate": 0, "cites": [], "content": "Mobility support has two aspects: user mobility and resource mobility. User mobility refers to how to automatically migrate the current program state and necessary data when the user moves from one edge node coverage domain to another so that the service handover is seamlessly. Currently, Cloudlet and CloudPath provide service migration through terminating/finishing existing task and starting a new VM/instance in the target edge node. Fine-grain migration and target edge node prediction are not supported. Resource mobility refers to i) how to dynamically discover and manage available resources, including both the long-term resources and short-term ones, and ii) when some edge devices are damaged how the system can be resumed as soon as possible with replacements. For example, PCloud and FocusStack supports edge devices dynamic join and leave with no task running. Intelligent dynamic resource management is still required for edge computing systems.", "id": "efe4a2f1-018b-4922-b876-e75becde3cd5", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "aa0b39df-82ae-4dab-a3c4-17e99ed6d2be", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ], [ "subsubsection", "Mobility Support" ] ], "subsections": [ "17c9b7d9-27e3-4b0f-a67c-bcd7775d13c6", "b64b53f2-9d6b-4d5d-90de-95bb60a6b2be", "36480d33-7cba-43e6-a451-cc5f81fa12fb", "4f1bfad2-fe7a-4d93-ac92-f95b4f346a5a", "3264a2b7-c71e-4be3-aad8-80db5b0b442f", "faa8d6ed-a0d3-4cfd-a483-16229a33a25b" ], "title": "Mobility Support" }, { "cite_extract_rate": 0, "cites": [], "content": "For edge devices with limited resources, how to ensure the fairness of multi-user usage, especially for the shared resources and rare resources. For example, a smartphone made up of various sensors and computing resources can act as an edge node to serve multiple users. However, as the smartphone has limited battery life, it is a problem to fairly allocate resources when receiving many requests. The resource competition is more intense, the requests are coming from different irrelevant tasks, it is hard to decide who should use the resource with only local information. Besides, the unfairness can be used as an attack strategy to make the critical task resource hungry, which may lead to main task failure. The existing edge computing systems do not pay much attention to the multi-user fairness, the basic solution (bottom-line) is to provide resource isolation, users only get what the edge node promised when it accepts the request. More fairness strategies require system support, like related task status updates, etc.", "id": "17c9b7d9-27e3-4b0f-a67c-bcd7775d13c6", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "efe4a2f1-018b-4922-b876-e75becde3cd5", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ], [ "subsubsection", "Mobility Support" ], [ "subsubsection", "Multi-user Fairness" ] ], "subsections": [], "title": "Multi-user Fairness" }, { "cite_extract_rate": 0, "cites": [], "content": "Unlike cloud computing, edge devices can be privately owned, such as gateway devices for smart home systems. When other users use such edge devices, obtain their data, and even take control of them, how to ensure the owner's privacy and guest users' data privacy is important. AirBox is a lightweight flexible edge function system, which leverages hardware security mechanism (e.g. Intel SGX) to enhance system security. Other system have not pay much attention to privacy and security. Existing cloud security approaches can be applied with the consideration of the resource limitation. Enhancing resource isolation, setting up privilege management and access control policies can be potential directions to solve the privacy problem.", "id": "b64b53f2-9d6b-4d5d-90de-95bb60a6b2be", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "efe4a2f1-018b-4922-b876-e75becde3cd5", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ], [ "subsubsection", "Mobility Support" ], [ "subsubsection", "Privacy Protection" ] ], "subsections": [], "title": "Privacy Protection" }, { "cite_extract_rate": 0, "cites": [], "content": "The system ultimately provides hardware interaction and basic services for upper-level applications. How to design interactive APIs, program deployment module, resource application and revocation, etc., are the key factors for the system to be widely used. Therefore, to design an edge computing system, we should think from an application developer's perspective. Specifically, to provide effective development and deployment services is a good idea to help improve the ecosystem of the edge computing system. For instances, EdgeX Foundry and Apache Edgent provide APIs to manage devices and analyze data respectively. ParaDrop supports monitoring devices and application status through developer APIs, and provides a web UI for users to manage/configure gateway as well as deploy applications to devices.", "id": "36480d33-7cba-43e6-a451-cc5f81fa12fb", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "efe4a2f1-018b-4922-b876-e75becde3cd5", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ], [ "subsubsection", "Mobility Support" ], [ "subsubsection", "Developer Friendliness" ] ], "subsections": [], "title": "Developer Friendliness" }, { "cite_extract_rate": 0, "cites": [], "content": "Edge computing involves multiple types of resources, each of which may belong to a different owner. For examples, the smart home gateways and sensors of the house owner, networks resources and base stations from the Internet service providers (ISP), traffic cameras owned by the government. How to access these resources and organize them according to the requirements of application and services, especially in emergency situations, is a problem that the edge computing system needs to consider. Existing researches assume we have the permission to use various devices belongs to different owners. In the initial stage, systems focus on the functionality perspective rather than implementation issues like price model. With the developing of edge computing, the real deployment issues should be solved before we enjoy all the benefits.", "id": "4f1bfad2-fe7a-4d93-ac92-f95b4f346a5a", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "efe4a2f1-018b-4922-b876-e75becde3cd5", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ], [ "subsubsection", "Mobility Support" ], [ "subsubsection", "Multi-domain Management and Cooperation" ] ], "subsections": [], "title": "Multi-domain Management and Cooperation" }, { "cite_extract_rate": 0, "cites": [], "content": "In cloud computing, the corresponding virtual machine can be allocated based on the resource requested by the user, and the cost model can be given according to the resource usage. In edge computing, an application may use resources from different owners. Thus, how to measure resource usage, calculate overall overhead, and give an appropriate pricing model are crucial problems when deploying an edge computing system. Generally, edge computing systems focus on how to satisfy the resource requirements of various services and applications, some of them pay more attention to the energy consumption due to the nature of mobile nodes, more complex overhead are not considered.", "id": "3264a2b7-c71e-4be3-aad8-80db5b0b442f", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "efe4a2f1-018b-4922-b876-e75becde3cd5", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ], [ "subsubsection", "Mobility Support" ], [ "subsubsection", "Cost Model" ] ], "subsections": [], "title": "Cost Model" }, { "cite_extract_rate": 0, "cites": [], "content": "Currently, specialized edge computing applications are still quite few. For examples, ParaDrop applications require additional XML configuration file to specify the resource usage requirements, SpanEdge needs developers to divide/configure tasks into local tasks and global tasks. Common applications are not directly supported to run in edge systems. How to automatically and transparently convert existing programs to the edge version and conveniently leverages the advantages of edge computing are still open problems. The compatibility should be considered in the edge system design. Specifically, how to adapt traditional applications to the new architecture and realize the basic functions so that they can run successfully, deserving further exploration and development.", "id": "faa8d6ed-a0d3-4cfd-a483-16229a33a25b", "level": "subsubsection", "origin_cites_number": 0, "parent_id": "efe4a2f1-018b-4922-b876-e75becde3cd5", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Key Design Issues" ], [ "subsubsection", "Mobility Support" ], [ "subsubsection", "Compatibility" ] ], "subsections": [], "title": "Compatibility" }, { "cite_extract_rate": 0, "cites": [], "content": "\\label{sec:conclusion}\nEdge computing is a new paradigm that jointly migrates the capability of networking, computation, and storage from the remote cloud to the user sides. Under the context of IoT and 5G, the vision of edge computing is promising in providing smarter services and applications with better user experiences. The recently proposed systems and tools on edge computing generally reduce the data processing and transmission overhead, and improve the efficiency and efficacy of mobile data analytics. Additionally, the integration of edge computing and deep learning techniques further fosters the research on edge-based intelligence services. This article introduced the representative systems and open source projects for edge computing, presented several energy efficiency enhancing strategies for performance consideration and technologies for deploying deep learning models at the edge, and suggested a few research directions during system design and analysis.\n\\nop{\n\\section*{Acknowledgment}\nThis work is partially supported by the National Key Research and Development Program of China (2016YFB1000302) and National Natural Science Foundation of China (61433019). Fang Liu and\nZhiping Cai are the corresponding authors.\n}\n\\bibliographystyle{IEEEtran}\n\\bibliography{reference}\n\\nop{\n\\begin{IEEEbiography}[{\\includegraphics[width=1in,height=1.25in,clip,keepaspectratio]{./fig/FangLiu.jpg}}]{Fang Liu}\nis an associate professor at the School of Data and Computer Science, Sun Yat-Sen University(SYSU), China. She received the B.S. and PhD. degrees in computer science from National University of Defense Technology, Changsha, China in 1999 and 2005, respectively. \nHer main research interests include computer architecture, edge computing and storage system.\n\\end{IEEEbiography}\n\\begin{IEEEbiography}[{\\includegraphics[width=1in,height=1.25in,clip,keepaspectratio]{./fig/GuomingTang.png}}]{Guoming Tang}\nis an assistant professor at the Key Laboratory of Science and Technology on Information System Engineering, National University of Defense Technology (NUDT), China. He received the Bachelor's and Master's degrees from NUDT in 2010 and 2012, respectively, and the Ph.D. degree in Computer Science from the University of Victoria, Canada, in 2017. Aided by machine learning and optimization techniques, his research focuses on computational sustainability in edge/cloud computing systems.\n\\end{IEEEbiography}\n\\begin{IEEEbiography}[{\\includegraphics[width=1in,height=1.25in,clip,keepaspectratio]{./fig/YouhuiziLi.jpg}}]{Youhuizi Li}\nis currently an assistant professor of Key Laboratory of Complex Systems Modeling and Simulation, Ministry of Education, Hangzhou. She is also with the School of Compute Science and Technology, Hangzhou Dianzi University, China. She received her B.E. and Ph.D. both in Computer Science from Xidian University in 2010, and Wayne State University in 2016, respectively. Her research interests include energy efficiency, edge computing and mobile system. She is a member of IEEE and CCF.\n\\end{IEEEbiography}\n\\begin{IEEEbiography}[{\\includegraphics[width=1in,height=1.25in,clip,keepaspectratio]{./fig/Zhiping.jpg}}]{Zhiping Cai}\nis a full professor in the College of Computer, NUDT. He received the B.Eng., M.A.Sc., and Ph.D. degrees in computer science and technology from the National University of Defense Technology (NUDT), China, in 1996, 2002, and 2005, respectively. His current research interests include network security and big data. He is a senior member of the CCF and a member of the IEEE. His doctoral dissertation has been rewarded with the Outstanding Dissertation Award of the Chinese PLA.\n\\end{IEEEbiography}\n\\begin{IEEEbiography}[{\\includegraphics[width=1in,height=1.25in,clip,keepaspectratio]{./fig/XingzhouZhang.jpg}}]{Xingzhou Zhang}\nis currently a Ph.D. candidate at State Key Laboratory of Computer Architecture, Institute of Computing Technology, Chinese Academy of Sciences, China. He received the Bachelor's degree from Shandong University in 2014. His main research interests include edge computing, machine learning, and computer systems.\n\\end{IEEEbiography}\n\\begin{IEEEbiography}[{\\includegraphics[width=1in,height=1.25in,clip,keepaspectratio]{./fig/Tongqing.jpg}}]{Tongqing Zhou}\nreceived the bachelor’s, and master’s degrees in Computer Science and Technology from National University of Defense Technology (NUDT), Changsha in 2012 and 2014, respectively. He is currently working toward the PhD degree in College of Computer, NUDT. His main research interests include wireless networks, mobile sensing, and network measurement.\n\\end{IEEEbiography}\n}\n\\end{document}", "id": "df3878a1-598b-4a96-97d4-ce6abf0afee7", "level": "section", "origin_cites_number": 0, "parent_id": "0055f39f-b56b-4a38-8ce6-f52d60700ea8", "prefix_titles": [ [ "title", "A Survey on Edge Computing Systems and Tools" ], [ "section", "Conclusions" ] ], "subsections": [], "title": "Conclusions" } ]
0
[ 8836, 4754, 3357, 7842, 3416, 2672, 2671, 7199, 4755, 4756 ]
1.586956
[ "David Ahmedt-Aristizabal", "Mohammad Ali Armin", "Simon Denman", "Clinton Fookes", "Lars Petersson" ]
A Survey on Graph-Based Deep Learning \\ for Computational Histopathology
2021
2021-07-01T07:50:35
cs.LG
" With the remarkable success of representation learning for prediction problems, we have witnessed (...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"01a5d650-4933-4e6e-b664-6a6a8f2896ea","level(...TRUNCATED)
1
[8845,6936,262,553,885,6938,6935,8305,550,6937,6940,6939,6943,6945,6942,6941,6944,8460,6948,6946,694(...TRUNCATED)
1.185043
["Jakob Gawlikowski","Cedrique Rovile Njieutcheu Tassi","Mohsin Ali","Jongseok Lee","Matthias Humt",(...TRUNCATED)
A Survey of Uncertainty in Deep Neural Networks
2021
2021-07-07T16:39:28
cs.LG
" Over the last decade, neural networks have reached almost every field of science and became a cruc(...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"039eb0dd-a257-4d85-a714-ad5fc65c6f0b","level(...TRUNCATED)
41
[4601,4598,4606,4612,4617,8806,7755,4596,4590,1828,3289,2441,4594,759,8807,4610,4600,4607,1044,3288,(...TRUNCATED)
0.976169
[ "Mokanarangan Thayaparan", "Marco Valentino", "André Freitas" ]
A Survey on Explainability in Machine Reading Comprehension
2020
2020-10-01T13:26:58
cs.CL
" This paper presents a systematic review of benchmarks and approaches for \\textit{explainability} (...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"03ca5aea-6a90-479f-934f-d9ab95896c40","level(...TRUNCATED)
42
[456,444,1798,439,460,4274,8761,4275,4277,2396,4276,4278,7802,7801,4279,7303,1147,8626,4281,4280,876(...TRUNCATED)
0.948755
[ "Saim Ghafoor", "Noureddine Boujnah", "Mubashir Husain Rehmani", "Alan Davy" ]
MAC Protocols for Terahertz Communication: A Comprehensive Survey
2019
2019-04-25T16:34:35
cs.NI
" Terahertz communication is emerging as a future technology to support Terabits per second link wit(...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"068ac4ba-0bb0-453c-9e5f-1071f02038e9","level(...TRUNCATED)
43
[]
null
[ "Liang Zhao" ]
Event Prediction in the Big Data Era: A Systematic Survey
2020
2020-07-19T23:24:52
cs.AI
" Events are occurrences in specific locations, time, and semantics that nontrivially impact either (...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"06b41427-1045-4fd8-b025-95b34ab2f969","level(...TRUNCATED)
44
[340,339,341,8342,344,343,166,342,8343,8344,9104,345,346,9105,347,59,9106,348,349,352,351,353,350,35(...TRUNCATED)
1.42435
[ "Rahul Mishra", "Hari Prabhat Gupta", "Tanima Dutta" ]
A Survey on Deep Neural Network Compression: Challenges, Overview, and Solutions
2020
2020-10-05T13:12:46
cs.LG
" Deep Neural Network (DNN) has gained unprecedented performance due to its automated feature extra(...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"06f26bac-bd30-4095-9869-55d16cc27805","level(...TRUNCATED)
45
[6312,6311,6313,688,6319,841,6315,504,4354,4628,8150,7634,8389,6318,6317,6314,6316,6320,9139,4351,63(...TRUNCATED)
0.751906
[ "Andrea Bandini", "José Zariffa" ]
Analysis of the hands in egocentric vision:\\ A survey
2019
2019-12-23T14:30:02
cs.CV
" Egocentric vision (a.k.a. first-person vision -- FPV) applications have thrived over the past few (...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"08af7ca6-1a67-48c9-9735-3113b8e18727","level(...TRUNCATED)
46
[4255,4254,4256,4257,4258,4259,4262,4261,3807,2571,4260,810,825,1766,514,4263,520,4265,206,2671,802,(...TRUNCATED)
1.064549
["Surangika Ranathunga","En-Shiun Annie Lee","Marjana Prifti Skenduli","Ravi Shekhar","Mehreen Alam"(...TRUNCATED)
Neural Machine Translation for Low-Resource Languages: A Survey
2021
2021-06-29T06:31:58
cs.CL
" Neural Machine Translation (NMT) has seen a tremendous spurt of growth in less than ten years, and(...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"109ef38f-b3f2-41dd-a765-4a141b2d87cf","level(...TRUNCATED)
47
[4965,303,2339,7873,2489,7874,4967,4968,4966,4969,4970,7875,4974,4973,4972,7200,4971,4975,168,38,498(...TRUNCATED)
0.746959
[ "Rui Wang", "Rose Yu" ]
Physics-Guided Deep Learning for Dynamical Systems: A Survey
2021
2021-07-02T20:59:03
cs.LG
" Modeling complex physical dynamics is a fundamental task in science and engineering. Traditional (...TRUNCATED)
[{"cite_extract_rate":0.0,"cites":[],"content":"","id":"11aa860a-a47f-4799-8633-7727b5a3edae","level(...TRUNCATED)
48
[6360,6356,6364,6355,6359,6363,6358,6362,6361,9004,6357,9003,6368,6370,9006,6369,6367,9005,6365,4714(...TRUNCATED)
0.85337
End of preview. Expand in Data Studio

SurGE

Welcome to the official Hugging Face repository for SurGE, a benchmark and dataset for end-to-end scientific survey generation in the computer science domain.

SurGE provides a comprehensive resource for evaluating automated survey generation systems through both a large-scale dataset and a fully automated evaluation framework.

More information at: https://github.com/oneal2000/SurGE

Overview

SurGE is designed to push the boundaries of automated survey generation by tackling the complex task of creating coherent, in-depth survey articles from a vast academic literature collection. Unlike traditional IR tasks focused solely on document retrieval, SurGE requires systems to:

  • Retrieve: Identify relevant academic articles from a corpus of over 1 million papers.
  • Organize: Construct a structured and hierarchical survey outline.
  • Synthesize: Generate a coherent narrative with proper citations, reflecting expert-authored surveys.

The benchmark includes 205 carefully curated ground truth surveys, each accompanied by detailed metadata and a corresponding hierarchical structure, along with an extensive literature knowledge base sourced primarily from arXiv.

Data Release

This repository contains all necessary components for working with the SurGE dataset:

Dataset Files & Formats:

  • Ground Truth Surveys: Each survey includes metadata fields such as title, authors, publication year, abstract, hierarchical structure, and citation lists.
  • Literature Knowledge Base: A corpus of 1,086,992 academic papers with key fields (e.g., title, authors, abstract, publication date, and category).
  • Auxiliary Mappings: Topic-to-publication mappings to support systematic survey generation.

The complete dataset can be downloaded from this Google Drive folder.

Then you will get the folder data .

Ground Truth Survey

A ground truth survey contains the full content of a survey and its citation information. However, due to space constraints, we cannot display it in its entirety here.

All ground truth surveys are available in data/surveys.json

A survey consists of the following fields:

Field Description
authors List of contributing researchers.
survey_title The title of the survey paper.
year The publication year of the survey.
date The exact timestamp of publication.
category Subject classification following the arXiv taxonomy.
abstract The abstract of the survey paper.
structure Hierarchical representation of the survey’s organization.
survey_id A unique identifier for the survey.
all_cites List of document IDs cited in the survey.
Bertopic_CD A diversity measure computed using BERTopic.

Literature Knowledge Base

The corpus containing all literature articles is available in: data/corpus.json

Example : Here, we present how articles are organized in the knowledge base. Overly long abstract has been appropriately shortened.

{
        "Title": "Information Geometry of Evolution of Neural Network Parameters While   Training",
        "Authors": [
            "Abhiram Anand Thiruthummal",
            "Eun-jin Kim",
            "Sergiy Shelyag"
        ],
        "Year": "2024",
        "Date": "2024-06-07T23:42:54Z",
        "Abstract": "Artificial neural networks (ANNs) are powerful tools capable of approximating any arbitrary mathematical function, but their interpretability remains limited...",
        "Category": "cs.LG",
        "doc_id": 1086990
    }

The following are explanations of each field:

Key Description
Title The title of the research paper.
Authors A list of contributing researchers.
Year The publication year of the paper.
Date The exact timestamp of the paper’s release.
Abstract The abstract of the paper.
Category The subject classification following the arXiv taxonomy.
doc_id A unique identifier assigned for reference and retrieval.

Auxiliary Mappings:

The mapping containing all queries and their corresponding articles is available in: data/queries.json

Each query in data/queries.json corresponds to a section or paragraph from the ground truth surveys with high citation extraction quality. The associated articles are the references cited in that part of the survey.

Below is an example, Overly long content has been appropriately shortened.

  {
        "original_id": "23870233-7f5b-4ef1-9d38-e6f3adb0fa48",
        "query_id": 486,
        "date": "2020-07-16T09:23:13Z",
        "year": "2020",
        "category": "cs.LG",
        "content": "}\n{\nMachine learning classifiers can perpetuate and amplify the existing systemic injustices in society . Hence, fairness is becoming another important topic. Traditionally...",
        "prefix_titles": [
            [
                "title",
                "Learning from Noisy Labels with Deep Neural Networks: A Survey"
            ],
            [
                "section",
                "Future Research Directions"
            ],
            [
                "subsection",
                "{Robust and Fair Training"
            ]
        ],
        "prefix_titles_query": "What are the future research directions for robust and fair training in the context of learning from noisy labels with deep neural networks?",
        "cites": [
            7771,
            4163,
            3899,
            8740,
            8739
        ],
        "cite_extract_rate": 0.8333333333333334,
        "origin_cites_number": 6
    }

The following are explanations of each field:

Key Description
original_id The identifier for the section where this query is from.
query_id The ID associated with the specific query.
content The content of the section.
prefix_titles A hierarchical list of titles of the section/subsection/paragraph
prefix_titles_query The question this passage is relevant to. The goal of the question is to retrieve relevant documents.
cites A list of document IDs that are cited within this section.
cite_extract_rate The ratio of extracted citations to the total number of citations in the original document.
origin_cites_number The total number of citations originally present in the section.

Note: Not every section in the ground truth surveys has a corresponding entry in queries.json. Only sections with a high citation extraction rate are included.
These entries can be used to train retrieval models, where prefix_titles_query serves as the query and cites contains the relevant document IDs.
The data in queries.json has not been pre-split into training and development sets—you may divide it manually as needed.

License

This project is licensed under MIT License. Please review the LICENSE file for more details.


license: mit task_categories: - text-generation language: - en

Downloads last month
10