Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
4 views9 pages

Zeitschriftenartikel 2024 SD Task

article

Uploaded by

sonia.benrejeb
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views9 pages

Zeitschriftenartikel 2024 SD Task

article

Uploaded by

sonia.benrejeb
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 9

OPEN CALL ARTICLE

Task-Oriented 6G Native-AI Network Architecture


Yang Yang, Jianjun Wu, Tianjiao Chen, Chenghui Peng, Jun Wang, Juan Deng, Xiaofeng Tao, Guangyi Liu, Wenjing Li, Li Yang, Yufeng He, Tingting Yang,
A. Hamid Aghvami, Frank Eliassen, Schahram Dustdar, Dusit Niyato, Wanfei Sun, Yang Xu, Yannan Yuan, Jiang Xie, Rongpeng Li, and Cuiqin Dai

Abstract aiming to provide connections for data trans-


mission, support user mobility, and ensure user
The vision for 6G networks is to offer per- experience. To achieve the 6G vision of per-
vasive intelligence and internet of intelligence, vasive intelligence and internet of intelligence,
in which the networks natively support artificial support for native AI at the 6G network archi-
intelligence (AI), empower smart applications tecture level is necessary [1], [2], [3]. Unlike
and scenarios in various fields, and create a the traditional communication services, AI is a
“ubiquitous-intelligence” world. In this vision, the data- and computing-intensive process, which
traditional session-oriented architecture cannot requires ubiquitous distribution, high real-time
achieve flexible per-user customization, ultimate performance, and high security and privacy—
performance, security and reliability required by aspects that 6G needs to support.
future AI services. To achieve the 6G vision of pervasive intelli-
In addition, users’ requirements for person- gence and internet of intelligence, the challenges
alized AI services may become a key feature and solutions have been provided in lots of articles
in the near future. However, the traditional AI [4], [5], [6], [7], they think the network archi-
deployment based on cloud/mobile edge tecture in 6G requires significant transformation
computing (MEC) has limitations such as low compared to traditional communications systems,
throughput, long delay, poor privacy, and high as normally the system architecture can only be
carbon emissions, resulting in the inability to pro- re-designed in the initial stage of one generation
vide personalized quality of experience (QoE) of radio network, and will keep unchanged in
assurance. By integrating AI in the network, the later releases. Such transformation will involve the
network AI has more advantages than cloud/MEC introduction of new resources—computing, data,
AI, such as better QoS assurance, lower latency, and algorithms—required by AI, and the design of
less transmission and computing overhead, and real-time management and control mechanisms
stronger security and privacy. Therefore, this arti- to support multi-node collaboration and heteroge-
cle proposes the task-oriented native-AI network neous resources collaboration, as well as security
architecture (TONA), to natively support the net- and privacy mechanisms for distributed AI work-
work AI. By introducing task control and quality flows across multiple nodes (including terminals
of AI services (QoAIS) assurance mechanisms at and network nodes).
the control layer of 6G, the TONA can achieve Based on the proceeding transformation, this
the finest service granularity at the task level for article further proposes a task-oriented native-AI
guaranteeing every user’s personalized QoE. network architecture (TONA) to meet personal-
ized AI service demand and requirements. This
Introduction article mainly:
The traditional communications system is 1. Introduces three-layer logical architecture of
session-oriented and typically provides connec- task management and control system, and
tions between specific terminals or between designs the task lifecycle management pro-
terminals and application servers. Its net- cedures, which include the collaboration of
work architecture offers a complete lifecycle multi-dimension heterogeneous resources
management mechanism (such as creation, (communication, computing, data, and algo-
modification, deletion, and anchor transfer of rithm) and multi-node at the control layer.
end-to-end (E2E) communication tunnels) and 2. Defines task-specific QoAIS indicators for
quality of service (QoS) assurance for sessions, the mapping from Service Level Agreement

Yang Yang (corresponding author) is with The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China, also
with the Peng Cheng Laboratory, Shenzhen, China, and also with the Terminus Group, Beijing, China; Jianjun Wu (corresponding author),
Chenghui Peng, and Jun Wang are with Huawei Technologies Company, Ltd., Shenzhen, China; Tianjiao Chen, Juan Deng, and Guangyi
Liu are with the China Mobile Research Institute, Beijing, China; Xiaofeng Tao is with the Peng Cheng Laboratory, Shenzhen, China,
and also with the Beijing University of Posts and Telecommunications, Beijing, China; Wenjing Li is with the Beijing University of Posts
and Telecommunications, Beijing, China; Li Yang is with ZTE Corporation, Nanjing, China; Yufeng He is with the Research Institute of
China Telecom, Beijing, China; Tingting Yang is with the Peng Cheng Laboratory, Shenzhen, China; A. Hamid Aghvami is with the King’s
College London, London, U.K.; Frank Eliassen is with the University of Oslo, Oslo, Norway; Schahram Dustdar is with TU Wien, Vienna, Digital Object Identifier:
Austria; Dusit Niyato is with Nanyang Technological University, Singapore; Wanfei Sun is with CICT Mobile Communication Technology 10.1109/MNET.2023.3321464
Company, Ltd., Beijing, China; Yang Xu is with Beijing OPPO Telecommunications Corporation, Ltd., Dongguan, China; Yannan Yuan Date of Current Version:
is with vivo Mobile Communication Company, Ltd., Dongguan, China; Jiang Xie is with the University of North Carolina at Charlotte, 18 April 2024
Charlotte, NC, USA; Rongpeng Li is with Zhejiang University, Hangzhou, China; Cuiqin Dai is with the Chongqing University of Posts and Date of Publication:
Telecommunications, Chongqing, China. 6 October 2023

IEEE Network Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore.
• January/February 2024 0890-8044/23©2023IEEE Restrictions apply.
219
algorithm) still requires cross-layer collaboration
It also provides data storage and processing functions as well as AI capabilities inside the network, in MEC. Consequently, the preceding problems
achieving higher security. involved in cloud AI remain unresolved.
Deploying AI functions (such as cloud and
MEC) at the application layer leads to low
(SLA) indicators—e.g., service requirement throughput, high latency, poor privacy, and high
zone (SRZ) and user satisfaction ratio carbon emissions. To address these problems,
(USR)—to QoAIS indicators, and discusses the network AI is launched to extend comput-
task-level QoS assurance to meet individual ing from the cloud to physically closer edges
requirements of different users. to end users. It also provides data storage and
3. Compares the network AI and cloud/mobile processing functions as well as AI capabilities
edge computing (MEC) in terms of QoAIS inside the network, achieving higher security.
indicators. Thanks to providing the AI exe- Although this “device-edge-cloud” architecture
cuting environments closer to UE, TONA is with edge cloud is expensive to deploy, it can
anticipated to have some advantages, such support compute-intensive, latency-sensitive,
as better data privacy protection, lower security-assured, and privacy-sensitive applica-
latency, and lower energy consumption. tions such as interactive virtual reality (VR) and
4. Lists some open issues, including distribut- augmented reality (AR) games, autonomous driv-
ed AI learning, mobility management, and ing, and smart manufacturing [7]. Therefore, it is
security assurance. becoming promising in various high-value-added
application scenarios.
Network AI By introducing AI in the network, 6G net-
The cloud AI architecture has been widely used work AI applies to three scenarios (shown in
in the 5G era to provide centralized computing, Fig. 1): Network element (NE) intelligence, net-
big data analysis, and AI training and inference work intelligence, and service intelligence. NE
services, where terminals provide data, mobile intelligence is the native intelligence of single
networks provide communication channels, and nodes, e.g., core network (CN) or radio access
clouds provide AI capabilities. Coordinating these network (RAN) nodes. Network intelligence
independent functions and resources among mul- refers to the collaboration of multiple intelli-
tiple facilities provides effective, flexible, smooth, gent NEs to achieve swarm intelligence. Both
and stable services and ensures QoE is extremely NE intelligence and network intelligence can
difficult. For latency-sensitive ultra-reliable low-la- be triggered internally or externally via open
tency communication (URLLC) services, MEC interface. Moreover service intelligence refers
deploys application servers close to base stations to the 6G network AI being provided as a ser-
and therefore has lower latency than cloud AI. vice, which is generally triggered by external
However, the AI platform is still deployed at the services and implemented in the network, with-
application layer. Joint optimization of connection out understanding the application service logic.
and AI resources (i.e., computing, data, model/ Put simply, NE intelligence and network intelli-
gence provide AI services for internal network
modules, and service intelligence provides AI
services for external third-party applications.
Here, we assume that some network units like
base stations and UEs will have some type of
AI processor which can be used for themselves
and the third parties.
To support the three scenarios, the 6G
native AI network architecture should have a
unified framework for different types of AI train-
ing and inference. For example, a distributed AI
environment must be built on the 6G network.
Specifically, the 6G native AI network architecture
must be able to: (1) use various native AI capa-
bilities (e.g., connection, computing, data, and
AI training and inference capabilities) of NEs and
terminals; (2) provide on-demand AI, computing,
and data services for networks and third-party
applications; and (3) guarantee the QoAIS in het-
erogeneous, dynamic, fully distributed, and other
complex wireless environments. This is the reason
that our proposed solution shift from a session-ori-
ented to a task-oriented architecture to address
the preceding challenges.

Network Paradigm Change


The TONA, as shown in Fig. 2, introduces the
orchestration and control functions as well as the
resource layer in network AI. The control function
uses control layer signaling to control multi-nodes
(UEs, base stations, and CN NEs) and heteroge-
FIGURE 1. Scenarios and requirements of 6G network AI. neous resources in real-time.

220 Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore. Restrictions apply.
IEEE Network • January/February 2024
FIGURE 2. Network paradigm changes.

We believe that the 6G network architec- models and reinforcement learning). Thus 6G
ture requires the following changes in the design networks need to introduce new resource man-
paradigm: agement mechanisms. However, it is difficult
1. Change 1: The object to be managed and to efficiently implement AI services on a single
controlled in network are changed from ses- node due to the bottlenecks in single-point com-
sions to tasks. puting, data privacy protection, and ultra-large
2. Change 2: The resources of the object model storage. Consequently, a new collabo-
are changed from one dimension to ration mechanism in 6G networks is required
multi-dimensions, from homogeneous to to implement computing, algorithm, and data
heterogeneous. collaboration among multiple nodes. Hence,
3. Change 3: The object control mecha- sessions and AI tasks have different technical
nism are changed from session-control to methods.
task-control. These differences show that the session-ori-
4. Change 4: The performance indicators of ented system cannot support native AI and that
the object are changed from session-QoS a new task-oriented system needs to be designed
to task-QoS. for the new resource management mechanism
and multi-node collaboration mechanism. This
Change 1: From Session to Task article defines a task that coordinates multi-
AI tasks differ from traditional sessions in terms of node and multi-dimensional resources at the 6G
technical objectives and methods. network layer to achieve a given objective. For
In terms of technical purposes, a traditional example, a federated learning in network needs
communications system provides session services, the coordination of multiple nodes of the base
typically between terminals or between terminals station and multiple UEs, and the coordination
and application servers, to transmit user data of communication, AI model, and computing
(including voice). Conversely, network AI (i.e., NE resources.
intelligence and network intelligence) aims to pro-
vide intelligent services for networks and improve Change 2: From Single-Dimension to Multi-Dimension
communication network efficiency. Service intel- Heterogeneous Resources
ligence seeks to provide app-specific intelligent The traditional wireless system establishes
services for third parties. Thus, sessions and AI tunnels and allocates radio resources for data
tasks have different purposes. transmission. Conversely, TONA implements
In terms of technical methods, to transmit collaboration among heterogeneous resources
user data, a traditional communications service of connection, computing, data and model/
needs to maintain a QoS assurance mechanism algorithm to execute AI tasks. Take an AI infer-
for user-oriented connection channels as well as ence task as an example. In this case, executors
their lifecycle management, such as E2E tunnels need to obtain certain resources to execute the
from UEs to base stations and then to the CN. tasks. Specifically, the executors need to obtain
This is necessary to provide QoS guarantee for computing resources like computing timeslots
the data transmission. Conversely, AI is a data- for the tasks, data resources like the data col-
and computing-intensive service. Compared lected in real-time or external data input, and
with sessions, AI introduces new resources, algorithm resources including a possible AI
including computing (e.g., CPU, GPU, and net- model such as a graph neural network (GNN),
work processing unit (NPU)), data (generated or a convolutional neural network (CNN), or rein-
used by AI), and algorithms (e.g., neural network forcement learning.

IEEE Network Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore.
• January/February 2024 Restrictions apply.
221
The QoS of traditional communication net-
Different AI scenarios have different requirements for AI service quality. works mainly considers connection-specific
performance indicators such as latency and
throughput of communication services [8].
Change 3: From Session-Control to Task-Control In addition to these traditional communication
Unlike session control, task management and resources, 6G networks will introduce new
control in network AI includes the following resources such as computing, algorithm, and
functions: (1) Decomposing and mapping from data, requiring an extension of evaluation indica-
external services to internal tasks, (2) Decompos- tors. At the same time, with the implementation
ing and mapping from service QoS to task QoS, of “Carbon Neutrality” and “Peak Carbon Diox-
and (3) Providing heterogeneous and multi-node ide Emissions” policies, the global AI industry’s
collaboration mechanisms to orchestrate and con- attention on data security and privacy, and users’
trol heterogeneous resources of multiple nodes at increasing requirements for network autonomy,
the infrastructure layer in real-time (to implement users will focus on more than just performance
distributed serial or parallel processing of tasks indicators in the future. The requirements on
and real-time QoS assurance). For a simple ser- aspects such as overhead, security, privacy, and
vice request, one service may correspond to or autonomy will increase, and these aspects will
be mapped as one task. For a complex service become new dimensions for evaluating QoS.
request (e.g., integration of multiple service flows, Consequently, the QoAIS indicator system needs
or a service flow with numerous calculations), one to be extended from the existing indicators during
service may be mapped to multiple nodes for sys- the initial design [9].
tematic execution. For example, the QoAIS indicators for AI train-
For function (3), the execution of an AI task ing services are as follows:
requires collaboration in two dimensions: 1. Efficiency: efficiency indicator boundary,
Heterogeneous Resources Collaboration: The training duration, generalization, reusabil-
execution of a task may require some or all of ity, robustness, explainability, consistency
the heterogeneous resources. For example, task between the loss function and optimization
deployment requires configuring the heteroge- objective, and fairness
neous resources, and task execution requires 2. Overhead: storage overhead, comput-
scheduling the heterogeneous resources in ing overhead, transmission overhead, and
real-time. power consumption
Multi-Node Collaboration: First, in a 3. Security: storage security, computing securi-
traditional communications network, connec- ty, and transmission security
tion-specific computing is mainly implemented 4. Privacy: data privacy, and algorithm privacy
on a single NE, and computing sharing and col- 5. Autonomy: fully autonomous, partially
laboration are not required. The emergence of autonomous, and manually controllable
AI is accompanied by large-scale AI training, QoAIS is an essential input for the network
large-model AI inference, and massive percep- AI orchestration and management system and
tual image processing, requiring significantly control functions. The orchestration and man-
more computing than traditional networks do. agement system decomposes and maps QoAIS
Simply expanding the computing capability of to generate QoS requirements of AI tasks, and
each NE across the entire network will result then maps the task QoS to QoS requirements
in high deployment costs. Hence, distributed of multi-dimensional heterogeneous resources.
computing is needed, which completes a task The management, control, and user plane mech-
collaboratively among multiple nodes through anisms are designed to ensure continuous QoAIS
shared computing. Second, as data owner- assurance.
ship awareness grows, data privacy protection
requirements become more stringent. For Architecture and Key Technologies
example, the raw data of User Equipment (UE) This section describes the logical architecture and
cannot be uploaded to networks for training. deployment options of TONA, and QoAIS details.
Federated learning solves this problem through
collaborative learning and gradient transfer at Logical Architecture of TONA
the data layer among multiple nodes. Third, First, we introduce fundamental basic concepts
model training consumes substantial computing in wireless network. A communications system
and storage resources to support native AI, and consists of a management domain and a con-
thus a good model needs to be shared within trol domain. The Operations Administration and
the network, and collaboration for models Maintenance (OAM) deployed in management
among multiple nodes is required. domain is used to operate and manage NEs
through non-real-time (usually within minutes)
Change 4: From Session-QoS to Task-QoS management plane signaling. The control domain
Unlike previous generations of mobile networks, is deployed on core network (CN) NEs, base sta-
6G networks are not just channels that serve tions, and terminals, and features with real-time
traditional communications services. Different controlling signaling (usually within milliseconds).
AI scenarios have different requirements for AI For example, an E2E tunnel for a voice call can
service quality. They demand an indicator mecha- be established within dozens of milliseconds by
nism to quantitatively or hierarchically convey user control signaling.
requirements while also orchestrating and con- Unlike the centralized, homogeneous, and sta-
trolling the comprehensive effect of AI resources. ble AI environment provided by the cloud, the
Therefore, this article proposes the quality of AI network AI faces the following technical chal-
service (QoAIS). lenges when embedded in the wireless networks:

222 Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore. Restrictions apply.
IEEE Network • January/February 2024
(1) AI needs to be distributed on numerous CN task scope and real-time task scheduling, and
NEs, base stations, and UEs. Therefore, it is nec- effectively manages the numerous, heteroge-
essary to consider how to manage the massive neous nodes and aware of dynamic change of
number of nodes efficiently in the architecture heterogeneous resources (e.g. channel status
design. (2) The computing, memory, data, and and computing load).
algorithm capabilities of different nodes vary sig- The following describes the detailed function-
nificantly, requiring the architecture design to also alities of TA, TS, and TE.
consider how to efficiently manage these het- TA manages the lifecycle of tasks (including
erogeneous nodes efficiently. (3) The dynamic deploying, starting, deleting, modifying, and mon-
variation of the channel status and the comput- itoring tasks) based on task QoS requirements. It
ing load need to be factored into the architecture also implements collaboration among heteroge-
design. neous resources to guarantee coarse-grained QoS
To address the aforementioned challenges, in the initial deployment phase.
TONA includes two logical functions, as TS controls and schedules tasks in the task
shown in Fig. 3: (1) AI orchestration and man- execution phase. It consists of the information
agement, called Network AI Management & collection and resource management modules.
Orchestration (NAMO); and (2) task control. Information collection requires that TS senses
NAMO decomposes and maps AI services to the computing load, data processing capabili-
tasks and orchestrates the AI service flows. It ties, algorithm models being used, and channel
is not performed in real-time and is generally conditions on a plurality of nodes in real-time.
deployed in the management domain. Task Based on this information, TS has a more real-
control introduces the Task Anchor (TA), Task time resource management capability than TA.
Scheduler (TS), and Task Executor (TE) func- For example, when the network environment
tions in the control domain in three layers. This changes, TS adjusts AI models and data pro-
layered design strikes a balance between the cessing functions or schedules connection and

FIGURE 3. Logical architecture of TONA.

IEEE Network Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore.
• January/February 2024 Restrictions apply.
223
FIGURE 4. Four deployment scenarios of TONA.

computing resources in real-time to achieve (CU) from the distributed units (DUs). In the latter
timely QoS assurance. mode, the CU may be deployed on the cloud for
TE is responsible for task execution and non-real-time signaling control and data transmis-
possible service data interaction. For example, sion. The DUs may be deployed closer to UEs for
federated learning needs to transfer intermediate real-time resource allocation, data transmission
gradient information among multiple nodes. and retransmission.
Scenario 1: gNodeB + UEs. In this scenario,
Deployment Architectures the gNodeB serves as both TA and TS, and the
The statuses of TEs (e.g., the CPU load, memory, UEs serve as TEs. Here, a UE is a computing
electricity, and UE channel status) change in real- provider and task executor, which accepts task
time. As such, deploying TA and TS close to each assignment and scheduling from the gNodeB. The
other can reduce the management delay. Accord- Uu interface and Radio Resource Control (RRC)
ing to the design logic of wireless networks, the layer between the gNodeB and the UE can be
CN and RAN need to be decoupled as much as enhanced to support task controlling and sched-
possible. For example, the CN should be inde- uling purposes.
pendent of RAN Radio Resource Management Scenario 2: CU + DUs. In this scenario, the
(RRM) and Radio Transmission Technology (RTT) CU serves as both TA and TS, and the DUs serve
algorithms. Therefore, this article recommends as TEs. Here, a DU is the computing provider and
that TA/TS be deployed on RAN and CN, named task executor. The F1 interface and F1-AP layer
RAN TA/TS and CN TA/TS, respectively. This way between the CU and the DU can be enhanced to
will allow TA to manage TEs in real-time flexibly. support task controlling and scheduling purposes.
Four deployment scenarios of TONA are shown Scenario 3: CU + DUs + UEs. In this scenario,
in Fig. 4 to describe the necessity and rationality the CU serves as TA, the DUs as TSs, and the UEs
of CN TA and RAN TA. These scenarios are only as TEs. Here, a UE is a computing provider and
examples—there may be other deployment sce- task executor, and the CU is the task manager. A
narios and architectures. DU observes a task allocated by the CU to UEs,
Assume that TA, TS, and TEs are deployed in and performs heterogeneous resources schedul-
RAN to perform federated learning between the ing and real-time QoS guarantee. This scenario
base station and UEs. Considering the 6G archi- separates TA from TSs. TSs are deployed lower
tecture is undetermined, this article reuses the 5G than TA is; TSs can therefore acquire the status
RAN architecture for reference. A gNodeB is a of TE heterogeneous resources more quickly to
5G base station, which can be deployed in stand- achieve real-time task QoS monitoring and rapid
alone mode or by separating the centralized unit adjustment of heterogeneous resources. The

224 Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore. Restrictions apply.
IEEE Network • January/February 2024
QoAIS Indicator Resource- specific Quantitative Indicator Non-quantitative Indicator
Performance indicator boundary, training Data Feature redundancy, integrity, data accuracy, and data Sample space balance, integrity, and
duration, generalization, reusability, preparation duration sample distribution dynamics
robustness, explainability, consistency with
Algorithm Performance indicator boundary, training duration, Robustness, reusability, generalization,
optimization objective, and fairness
convergence, and optimization objective matching degree explainability, and fairness
Computing Computing precision, duration, and efficiency None
Connection Bandwidth and jitter, delay and jitter, bit error rate and None
jitter, and reliability
TABLE 1. Mapping between QoAIS indicators and resource QoS indicators in the AI training service.

Uu interface and RRC/Medium Access Control robustness, reusability, generalization, and


(MAC) layer between the CU/DU and the UE explainability) cannot be evaluated quantitatively.
can be enhanced to support task controlling and Therefore, we need to consider indicators that
scheduling purposes. reflect user requirements and introduce them by
Scenario 4: CN + gNodeB + UEs. In this sce- phase.
nario, the CN serves as TA, the gNodeB serves
as TS, and the UEs serve as TEs. Here, a UE is the Task Procedures
computing provider and task executor. The Non From the E2E procedure perspective, NAMO sub-
Access Stratum (NAS) interface and NAS layer mits the AI service request to TA for execution
between the CN and the UE can be enhanced after receiving an external service request. The
to support task controlling purposes, and the E2E procedure is as follows:
Uu interface and RRC/MAC layer between the 1. Generate or import an AI use case, which is
gNodeB and the UE can be enhanced to support an AI service request submitted by a user to
task scheduling purposes. the network. This use case may call one or
In this example, TA, TS, and TE are only logical more types of network AI services, such as
functions, which may be deployed on the same or AI training, verification, and inference.
different nodes depending on the scenarios. Log- 2. Decompose the use case into one or more
ically, a single node may have multiple functions AI services.
(any combination of TA, TS, and TE). 3. Decompose an AI service into one or more
AI tasks (AITs), and decompose the AI ser-
Task QoS Assurance vice QoAIS into the AI task QoS.
To guarantee QoAIS, the aforementioned hier- 4. Determine the anchor position of an AIT.
archical management and control architecture 5. Decompose the task QoS into resource
is implemented through three-layer closed-loop QoS requirements, and specify the hetero-
management. The TS layer monitors and opti- geneous resources required by the AIT.
mizes the heterogeneous resources in real-time 6. Determine and configure the heteroge-
to ensure task QoS within TA resource configu- neous resources required by the AIT. This
rations. When the task QoS guarantee is beyond involves selecting nodes (those that partic-
the range of the TS layer (e.g. the computing ipate in computing and provide data and
resource controlled by the TS is not sufficient for a algorithms/models), establishing connec-
task, the TS should report to TA to allocate more tions between nodes, and updating the
computing resource to guarantee the QoS of this configurations.
task), the TA layer modifies the overall resource 7. Among the selected nodes, determine and
configuration. For example, the TA layer adjusts adjust the computing allocation in real-
the network nodes involved in the task and time, optimize the communication connec-
replaces the model or data warehouse. When the tion quality, collect the required data, and
task QoS guarantee is beyond the range of the TA replace or optimize the algorithms/models.
layer, NAMO performs optimization by changing This is necessary to ensure the task QoS
the anchor position of an AI task or decomposing and further guarantee the QoAIS.
the mapping between AI services and AI tasks. The management layer has poor real-time per-
Table 1 lists the mapping between the QoAIS formance. Although it can obtain a wide range
and resource QoS indicators. The QoAIS indica- of network information, such information is
tors are decomposed into task QoS indicators and coarse-grained. Furthermore, the management
then mapped to resource QoS indicators, which layer cannot obtain real-time information about
jointly guarantee the QoAIS at the management, radio links and terminal resources. Conversely,
control, and user planes. The QoS indicators in the control layer has good real-time perfor-
each resource dimension are classified into indi- mance. However, while it can obtain accurate
cators suitable for quantitative evaluation (such information, the range of this information is lim-
as resource overheads) and qualitative evalu- ited. Hence, some functions are suitable for the
ation (such as security level, privacy level, and management or the control layer, and others may
autonomy level). For the indicators suitable for be achieved through the collaboration of both
quantitative evaluation, the quantization solutions layers.
are mature or easy to formulate, such as train-
ing duration, algorithm performance boundary, Advantage Analysis
computing precision, and resource overheads. Compared with cloud/MEC AI, the TONA
However, some other indicators (e.g., model and QoAIS have the following advantages

IEEE Network Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore.
• January/February 2024 Restrictions apply.
225
TONA, during task execution, TONA can detect
TONA has native data security and privacy protection capabilities because it processes data inside the dynamic changes of computing loads on termi-
the network. nals in real-time and promptly adjust the computing
resource allocation, calculation precision, and serial
or parallel computing mode on the network. As
such, this approach can ensure QoAIS for custom-
ized computing offloading services.
Cloud/MEC AI TONA
QoAIS Difficulty to guarantee the personalized QoAIS. Easy to guarantee the personalized QoAIS. Latency
TONA computing is distributed on NEs closer
Latency Higher latency due to out-network processing (e.g. Lower latency due to in-network processing to UEs or even directly on UEs to process data
second/minute level). (e.g. millisecond level). locally. This not only successfully achieves
Resource overhead Larger transmission overheads. Less transmission overheads. real-time and low-latency AI services, but also sig-
nificantly reduces data transmission. In the cloud/
Larger computation overheads. Less computation overheads. MEC AI mode, a large amount of data needs to
be transmitted to the cloud/MEC for training,
Security Data privacy is ensured by the application layer. Native security and privacy via in-network processing.
meaning that E2E data transmission takes longer
TABLE 2. Performance comparison of cloud/MEC AI and TONA. to complete.
Take joint device-cloud AI inference as an
example [14], [15]. Cloud/MEC AI transmits data
(summarized in Table 2) in meeting users’ custom- from devices to the cloud, performs real-time
ized AI service requirements: training/interference, and transmits the results
back to devices. The long transmission distance
QoAIS Assurance causes high latency, making it difficult to meet the
Dynamic wireless environments require joint requirements of ultra-low latency scenarios such
optimization of the heterogeneous resources as Industrial Internet of Things (IIoT), even if the
(connection and three AI resources) to achieve application server is deployed on the MEC. By
precise QoAIS assurance. In TONA, all hetero- contrast, for TONA, data processing is terminated
geneous resources are inside the network and within a network, the E2E transmission latency is
can perceive each other. Furthermore, a real-time as low as 1 millisecond, enabling ultra-low latency.
(within milliseconds) collaboration mechanism
is designed at the control layer. Conversely, the Overhead
cloud/MEC AI lacks a collaboration mechanism TONA can optimally allocate resources through
between the communication resources and the the real-time collaboration mechanism of the
three AI resources, meaning that these resources heterogeneous resources, maximizing the overall
cannot observe each other in real-time. Generally, resource utilization and reducing the transmission
they observe each other through the management and computing overheads. Conversely, because
layer (with non-real-time capability openness) or the cloud/MEC AI cannot adapt to dynamic envi-
the application layer (within seconds or minutes), ronments, it allocates resources based on only
which cannot adapt to dynamic wireless environ- the maximum resource consumption to ensure
ment changes in real-time, and cannot guarantee QoAIS. As a result, the overall resource utilization
QoAIS. is low, and the resource overhead is high.
Take device-cloud joint AI training as an exam- Take joint device-cloud AI training as an
ple [10], [11], the Cloud AI cannot be aware of example. For Cloud/MEC AI, long device-cloud
the connection’s real-time status to adjust the het- distance causes large transmission overheads.
erogeneous resources and thus cannot provide On the other hand, for TONA, data is processed
customized training solutions for users with dif- nearby, effectively reducing data transmission
ferent connection performances. Meanwhile, for overheads.
TONA, the network detects environment changes Furthermore, Cloud/MEC AI cannot measure
(such as terminal movement, disconnection, and quality of wireless connections in real-time. Dif-
burst interference) in real-time and quickly adjusts ferent connections status of TEs lead to low AI
the joint training solution. For example, in TONA efficiency and increase computing overhead. In
the network can change the split learning [12], federated learning, for example, straggler ter-
[13] point to reduce the intermediate data size minals may be abruptly disconnected from the
when the UE is far away from the base station. network or cause a long delay. If a large amount
Thus, QoAIS can be achieved for customized AI of straggler data is discarded, the number of
training services. training samples is reduced, affecting the conver-
Take device-cloud computing offloading as an gence efficiency of the current round. If a long
example. If a terminal’s local computing resource delay occurs, the iteration time of the current
does not meet the requirements of computing-in- round is prolonged. However, for TONA, the net-
tensive services, cloud/MEC AI offloads some work can detect each UE’s channel status and
computation to the cloud. During the execution set a longer local training period for the straggler
of computing tasks, the computing resource uti- to reduce the total reporting numbers when the
lization of terminals changes in real-time (within UE‘s data rate is low. This improves the overall
milliseconds). The non-real-time collaboration of AI training efficiency and reduces the computing
cloud AI (within seconds or minutes) cannot trace overhead.
users’ computing requirements in real-time, nor
can it promptly offload computing to the cloud. Security
As such, this approach fails to meet users’ custom- TONA has native data security and privacy pro-
ized QoAIS requirements. On the other hand, for tection capabilities because it processes data

226 Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore. Restrictions apply.
IEEE Network • January/February 2024
inside the network. Unlike TONA, the cloud/MEC task concept and TONA proposed in this article
AI protects data privacy only at the application support not only AI tasks, but also sensing-, com-
layer. puting- and data processing-specific tasks.

Open Research Issues Acknowledgment


Although the industry has reached a preliminary The work of Yang Yang was supported in part by
consensus on 6G native AI networks, some effi- the National Key Research and Development Pro-
cient support and standardization methods need gram of China under Grant 2020YFB2104300.
further research and development. Special thanks should be given to 6GANA, which
1. Distributed AI Learning: Distributed AI has fully discussed most of the technical view-
learning involves collaboration and inter- points and solutions mentioned in this article.
action among multiple TEs. The definition
of the collaboration mechanisms and inter- References
action information vary according to algo- [1] Wen Tong and Peiying Zhu, 6G: The Next Horizon: From
rithms. Therefore, we need to study how to Connected People and Things to Connected Intelligence.
natively and efficiently support distributed Cambridge, U.K.: Cambridge Univ. Press, 2021. [Online].
AI from the architectural perspective. Available: https://www.vitalsource.com/en-ca/products/6g-
the-next-horizon-v9781108997287
2. Mobility: Assuming a UE participates in the [2] 6GANA WhiteArticle. (May 31, 2021). From Cloud AI to
task process (e.g., the UE is a participant in Network AI: A View from 6GANA. [Online]. Available:
task execution), when the UE moves from https://www.6g-ana.com/upload/file/20211011/63769566
one base station to another, how to guar- 58802608005614761.pdf
[3] G. Liu et al., “The SOLIDS 6G mobile network architecture:
antee the service continuity for AI tasks Driving forces, features, and functional topology,” Engineer-
is a critical problem (e.g., how to achieve ing, vol. 8, no. 1, pp. 42–59, 2021.
zero-millisecond interruption). [4] Q. Tang et al., “Internet of Intelligence: A survey on the enabling
3. Security Assurance: Considering the dis- technologies, applications, and challenges,” IEEE Commun.
Surveys Tuts., vol. 24, no. 3, pp. 1394–1434, 3rd Quart., 2022.
tributed deployment of TEs in TONA, the [5] Y. Yang et al., Intelligent IoT for the Digital World. Hoboken,
access of multiple heterogeneous devices NJ, USA: Wiley, 2021.
and distributed AI learning pose significant [6] Y. Yang, “Multi-tier computing networks for intelligent IoT,”
challenges to network AI security. There- Nature Electron., vol. 2, pp. 4–5, Jan. 2019.
[7] Y. Yang et al., “6G network AI architecture for everyone-cen-
fore, for the distributed communication and tric customized services,” 2022, arXiv:2205.09944.
learning of TEs, the implementation of port [8] Quality of Service (QoS) Concept and Architecture, 3GPP
monitoring, privacy protection and security document TS23.107, Version 16.0.0, 2010.
isolation is an important research direction [9] U. Mikko, et al., “6G vision, value, use cases and technol-
ogies from European 6G flagship project Hexa-X,” IEEE
to ensure the security of TONA. Access, vol. 9, pp. 160004–160020, 2021.
Conclusion [10] Z. Zhou et al., “Edge intelligence: Paving the last mile of
artificial intelligence with edge computing,” Proc. IEEE, vol.
To meet the 6G vision of pervasive intelligence 107, no. 8, pp. 1738–1762, Aug. 2019.
[11] Y. Xiao et al., “Toward self-learning edge intelligence in 6G,”
and internet of intelligence, TONA is proposed to IEEE Commun. Mag., vol. 58, no. 12, pp. 34–40, 2020.
support efficient collaboration of heterogeneous [12] C. Thapa et al., “SplitFed: When federated learning meets
resources and multi-node in wireless networks, split learning,” in Proc. AAAI Conf. Artif. Intell., 2022, vol. 36,
and to provide new services in the form of tasks at no. 8, pp. 8485–8493.
[13] P. Vepakomma et al., “Split learning for health: Distribut-
the network layer. By bringing new dimensions of ed deep learning without sharing raw patient data,” 2018,
resources to 6G networks (i.e., computing, data, arXiv:1812.00564.
and model/algorithm), this architecture enables [14] K. B. Letaief et al., “Edge artificial intelligence for 6G: Vision,
the SLA assurance of computing related services enabling technologies, and applications,” IEEE J. Sel. Areas
Commun., vol. 40, no. 1, pp. 5–36, Jan. 2022.
such as AI services, further explores the appli- [15] J. Shao and J. Zhang, “Communication-computation trade-
cation scenarios of 6G networks, and enriches off in resource-constrained edge inference,” IEEE Commun.
the value of wireless networks. Furthermore, the Mag., vol. 58, no. 12, pp. 20–26, Dec. 2020.

IEEE Network Authorized licensed use limited to: TU Wien Bibliothek. Downloaded on June 11,2024 at 12:38:23 UTC from IEEE Xplore.
• January/February 2024 Restrictions apply.
227

You might also like