Abstract
The popularity of smart mobile devices has led to a tremendous increase in mobile traffic, which has put a considerable strain on the fifth generation of mobile communication networks (5G). Among the three application scenarios covered by 5G, ultra-high reliability and ultra-low latency (uRLLC) communication can best be realized with the assistance of artificial intelligence. For a combined 5G, edge computing and IoT-Cloud (a platform that integrates the Internet of Things and cloud) in particular, there remains many challenges to meet the uRLLC latency and reliability requirements despite a tremendous effort to develop smart data-driven methods. Therefore, this paper mainly focuses on artificial intelligence for controlling mobile-traffic flow. In our approach, we first develop a traffic-flow prediction algorithm that is based on long short-term memory (LSTM) with an attention mechanism to train mobile-traffic data in single-site mode. The algorithm is capable of effectively predicting the peak value of the traffic flow. For a multi-site case, we present an intelligent IoT-based mobile traffic prediction-and-control architecture capable of dynamically dispatching communication and computing resources. In our experiments, we demonstrate the effectiveness of the proposed scheme in reducing communication latency and its impact on lowering packet-loss ratio. Finally, we present future work and discuss some of the open issues.
Keywords: 5G, artificial intelligence, LSTM, mobile traffic, uRLLC
1. Introduction
The fifth generation of mobile communication networks (5G) with leveraging real-time artificial intelligence (AI) is a promising prospect towards fulfilling the high-level requirements of the heterogeneous Internet of Things (IoT). The 3rd Generation Partnership Project (3GPP) defines three emerging application scenarios: enhance mobile broadband (eMBB), massive machine-type communication (mMTC), and uRLLC [1]. While eMBB focuses on high spectral efficiency, uRLLC, has been recognized as a key trend of 5G, but faces tough challenges in order to meet the stringent requirements, such as high reliability and low latency. As uRLLC is becoming increasingly prevalent, exploiting new technologies such as AI, would be essential to tackle more complex tasks in the presence of high traffic flow. This trend is driven by the ever increasing popularity of smart devices, which are having a huge impact on the trade-offs between the services of telecommunications suppliers and users’ demands [2]. For instance, user’s quality of experience (QoE) is an essential prerequisite for providing intelligence and personalized services for real-time interactions. Within the integrated framework of the IoT and the cloud (IoT-Cloud), the anticipated increase in mobile-traffic flow can also have a profound effect on computation intensity, as well as dispatching pressure on the edge cloud (base-station) and remote cloud (data center) [3–4].
In addition, the intensified mobile-traffic flow can cause a shortage of computing and networking resources in which case applications might not be able to respond to users’ requests in a timely manner. Another important issue is the allocation of bandwidth resources for heterogeneous (cloud) services [5]. For instance, the proliferation of resources to support intelligent services and the gradual transition from traditional to heterogeneous IoT, creates conflicting demands for network operators and service providers. For the IoT-cloud in particular, existing data-driven methods directly unload computing tasks without being analyzed, processed, and controlled by base-stations. These can greatly undermine transmission reliability, as well as impact communication latency for delay sensitive applications [6]. Since such a high latency cannot be tolerated by uRLLC users, the main challenge is how to efficiently analyze and control the mobile-traffic-flow data in order to achieve communications with ultra-low latency and ultra-high reliability.
With the anticipated integration of 5G networks, edge computing and the IoT-Cloud, AI assisted mobile-traffic-flow, prediction, and management can offer a viable solution for future mobile-network planning and dynamic resource allocation. In view of recent advances in smart mobile devices, base-stations, and remote clouds, together with the latest development of computing and storage techniques, AI can play a significant role in supporting uRLLC scenarios in order to meet its strict requirements in terms of latency and reliability. Depending on advanced machine-learning (ML) methods, AI is transforming from traditional pattern recognition to the management of complex systems. In the past few decades, ML rose and fell several times as the main branch of AI. It has now reached sufficient maturity that it can be permeated into the design of many complex systems, including wireless communications [7]. If algorithms, such as LSTM [8–9], are deployed at different positions in mobile networks, mobile-traffic flow can be predicted intelligently. This is because LSTM is suitable for processing and predicting important events with relatively long intervals in the time series [10].
Therefore, in our proposed frame work, an LSTM-based deep-learning algorithm is considered to predict the uRLLC mobile-traffic flow received by a single edge cloud. The predicted peak value, representing the traffic of the entire network at each time instant, is sent to a remote cloud. At the remote cloud, resources are dispatched and allocated dynamically based on traffic adaptation using a cognitive engine and an intelligent mobile traffic module to balance the network load. This contributes towards achieving high reliability and low latency of communications, hence enhancing the users’ QoE.
This paper is organized as follows. In Section 2, after a brief description of the three application scenarios of 5G, we present an IoT-Cloud architecture focusing on the uRLLC application scenario. Then, a description of the mobile-traffic flow is given to disclose the interactions between user devices, edge clouds, and remote clouds. Section 3 proposes a uRLLC mobile-traffic flow prediction algorithm that exploits an LSTM-base algorithm operating in a single-site mode. In Section 4, we present a novel intelligent mobile-traffic control architecture for a large-scale multi-site IoT-Cloud. Experiments and performance evaluations of the proposed mobile-traffic prediction and control framework are given in Section 5. Section 6 discusses some open issues about users’ mobility prediction, strategy sharing and risk perception. Finally, Section 7 summarizes the study.
2. A heterogeneous Internet of Things based on uRLLC
2.1. The three application scenarios of 5G
As mentioned before, the three core services of 5G, which have been defined by 3GPP are: eMBB, mMTC and uRLLC [1]. The applications covered by these scenarios are shown in Fig. 1. The aim of eMBB is to enhance users’ QoE based on existing mobile broadband business scenarios. It mainly attempts to achieve an ideal short distance and personal communication. The main use case scenarios of eMBB are smart homes, VR/AR, smart devices, smart buildings, etc. Mobile broadband businesses with large traffic flows, such as 3D/UHD video, need the support of broadband resources to guarantee smooth transmission of the mobile-data [11].
Fig. 1.
Application scenarios of eMMB, mMTC and uRLLC
The mMTC is intended to support IoT on a large scale radio coverage among humans and objects. Its main features are low cost, low energy consumption, short packets with small quantity of data, and a large number of connections. The mMTC supports applications such as the agricultural environment, smart metering, smart cities, logistics, tracking, massive information interaction, etc.
Finally, uRLLC aims to improve users’ QoE [12] by establishing a highly reliable and stable communications. At the same time, uRLLC has to meet tough requirements with respect to network latency. The main application scenarios of uRLLC are smart factories, autonomous automobiles, remote surgery, affective interaction, etc. These application scenarios require reliable communications among terminal equipment, base-stations, and the cloud server, as well as intelligent computing and dynamic resource allocation.
In this paper, we mainly concern the prediction and control of mobile-traffic flow to support uRLLC services in terms of high reliability, low latency, and extremely high usability. Bear in mind that except for some regulated daily log data transmissions, uRLLC users’ service requests on information transmissions are generally unpredictable and would require an advanced knowledge of traffic flow in order to efficiently execute resource-allocations and task computations. Due to the advantages of time-series prediction, LSTM is a natural approach to solving the problem of mobile-traffic-flow prediction in uRLLC scenarios. The algorithm can be deployed in edge clouds and remote clouds.
2.2. IoT-Cloud architecture based on uRLLC
Our proposed IoT-cloud architecture for uRLLC scenario is shown in Fig. 2. Further details of the architecture are described as follows:
Fig. 2.
IoT-Cloud architecture based on uRLLC
The user-device layer mainly covers common IoT scenarios aimed at uRLLC (details in Section 2.1) where each smart device requests a uRLLC-based service with an uncertain type and volume of data which would be uploaded to the edge cloud via a wireless network.
The edge-cloud layer contains wireless connectors such as base-stations, access points, and routers. The edge cloud, which provides wireless access to each IoT device, performs lightweight caching and computing as well as unloading complex computing tasks to the remote cloud.
The remote-cloud layer usually contains SDN [13–14] controllers and a central node with the deployment of cognitive intelligence engine [15]. The available resources, with the help of data-cognitive engine on the cloud server, are utilized to perform mobile-traffic-flow prediction and dynamic resource allocation. The internal modules and functionality of the cognitive engine will be introduced in Section 4.
Traditional traffic prediction algorithms, such as short-term correlation prediction methods (Markov model, autoregressive model (AR) and its variants (ARMA, ARIMA)), mainly focus on modeling of stationary flow characteristics. However, these algorithms require high self-similarity of traffic time series. For example, AR algorithm adopts linear mapping method which uses p-order autoregressive model . Thereinto, the current flow value of Vt is expressed as the sum of a linear combination of one or more historical flow values of Vt–i, a constant term of c and a random error of εt. Although AR algorithm needs little data, the dependence on its own variable series easily leads to inaccurate prediction results. Therefore, the extended ARIMA algorithm adds a sliding average coefficient of q, a difference number of times of d, and a lag operator of L. . However, ARIMA algorithm does not consider other relevant variables in the process of modeling which is vulnerable to the impact of network dynamic changes.
In other words, existing traffic prediction and network optimization methods mainly face the following three difficulties: 1) Human intervention, traditional network optimization modelling is constructed by experts with domain knowledge. Such knowledge-driven model is costly and inefficient in implementation, 2) invalid model. More users, more access methods, more complex functions and network entities make the mathematical model and relationship functions constructed in advance unable to match the reality, 3) high complexity. From the history of prediction algorithms, we can see that researchers often add more variables to the relationship function to improve the effect of dynamic prediction, but at expense of multidimensional computing complexity.
In fact, the above problems can be solved by balancing the mobile traffic data and reducing bandwidth occupancy in the same period. Such a reduction can be achieved by efficiently predicting mobile-traffic flow and assessing the priority of the users’ requests in transmission queues with dynamically allocating the resources for communication and computing. Therefore, under the proposed IoT-cloud architecture, our main objective is to develop an efficient machine-learning-based prediction and control algorithm that can ensure high reliability and low latency as required for uRLLC communications.
3. Mobile-traffic-flow prediction based on LSTM
In the uRLLC scenario, information (in terms of service requests and data communications) is produced in large quantities. This information is transmitted to the edge cloud via wireless channels in the form of packets. Bear in mind that the network’s bandwidth are generally limited and if the data from users is too excessive at certain time intervals, the network will block any transmissions without any alternative. This could consequently undermine the communication integrity due to packet loss. A viable approach to improve the communication performance would to develop an efficient prediction scheme that can dynamically manage the mobile-traffic flow.
We use LSTM algorithm to build a uRLLC mobile-traffic-flow prediction model of a single edge cloud [8–9], as shown in Fig. 3. Furthermore, our proposed prediction module is made up of several stacks of attention-based LSTM layers. Each layer operates as the corresponding part of the basic LSTM cell with a attention mechanism.
Fig. 3.
uRLLC mobile-traffic-flow prediction of single-site based on LSTM
3.1. LSTM cell
First, within each time interval, we extract the peak value from the actual traffic-flow wave serial data, i.e., the input data set, V = {V1, V2, …, Vt}, where Vt is the peak mobile-traffic flow at the current time slot tc. We seek to predict the peak mobile-traffic flow Vt+1 at the next time instant tc+1. The solution process of a single LSTM cell is presented in the following formula (1) [10].
| (1) |
where, ft, Ct, ht signify the forgetting gate, the input gate, and the output gate, respectively. Wf, WC bf and bC represent the weights and the bias factors of the forgetting gate and the input gate, respectively. Furthermore, σ and tanh are activation functions. The flow feature ht is extracted from a single LSTM model.
3.2. Attention model
An attention model is widely used in natural language processing (NLP) area [16]. It is used to improve the neural network machine translation (NMT) as a sequence to sequence (encoder to decoder) model as in the case of the human brain, which pays different attentions to different events. In this section, we introduce the attention mechanism to improve the weight of the peak value in mobile traffic flow prediction. Such a mechanism aims to exploit any correlation between some key input features and output values.
Assume the output set of ith LSTM layer is , where i ∈ {1,2, …, L}. Therefore, the intermediate code κ = hi–1 represents the output state of a series LSTM cells in (i−1)th layer, where . F(·) indicates the LSTM process calculated by formula (1). When is fed into the attention model, the intermediate code κ is used to conduct a tanh process to obtain the aggregation state , as shown below:
| (2) |
where, Wκm and Whm are the weights of the above two inputs. In order to get the highest value for , we compute weights by a softmax function as:
| (3) |
As becomes larger, converges more towards . Then, we define as the mapping of in learning direction as shown below:
| (4) |
where, represents the transfer of matrix set [Wκm, Whm]. It should be noted that the value of tends to increase at high peak time, while dropping at a lower peak time. Thus, the output of the attention model can be defined as:
| (5) |
Finally, we use the combination of the attention model output and the LSTM cell output from the previous time as the input data for the current time slot, i.e., , where such a process will be loop execution until i = L.
3.3. Output layer
A fully connected layer synthesizes the above-extracted features in Lth layer to obtain the output sequence P = {P2, P3, …, Pt+1}, where Pt+1 corresponds to the predicted value of the mobile-traffic flow at the next time slot, t + 1. By using a combination of the peak flow values from the previous time slots, we can then predict the value for every moment of mobile-traffic flow in a time span (i.e., the predicted traffic-flow wave).
To maximize the accuracy rate of the prediction (i.e., minimizing the root-mean-square error, which is used to measure the deviation between the observed value and the true value), the network should be able to accurately allocate the communication resources in order to prevent congestion according to the following formula (6).
| (6) |
Algorithm 1 shows the detailed training process of our intelligent mobile-traffic prediction using the proposed attention-based LSTM algorithm. The storage and computing capacity of edge clouds (i.e., base-stations or wireless access points) are usually insufficient, so a smaller and more accurate learning model is more suitable for high-reliability and low-latency services. Compared with the existing works, [8] uses a single LSTM cell to conduct prediction and [9] doesn’t include multi-layer mechanism.
Algorithm 1.
Training process of mobile-traffic prediction.
| Input: | |
| Input Real traffic flow V; | |
| LSTM layers L | |
| Memory time slots n_input | |
| Number of iterations epoch | |
| Quantity of data input for each time batch_size. | |
| Output: | |
| 1: | Data normalization |
| 2: | Initial hidden-layer output h and LSTM state o |
| 3: | Create training set and reshape input data |
| 4: | LSTM_Cell ← formula(1) |
| 5: | Repeat |
| 6: | for i = l in range(L) |
| 7: | for step in range(n_input × batch_size) |
| 8: | ht, ot ← LSTM_Cell(input_Xt−1, ot−1) |
| 9: | if i ≠ L |
| 10: | zt ← formula(2 ∼ 5) |
| 11: | input_Xt ← Vt ← concat(ht−1, zt) |
| 12: | If Loss(ht+1, input_Yt) ∝ 0 |
| 13: | |
| 14: | End if |
| 15: | Until ← formula(6) |
| 16: | |
Thereinto, L can be dynamically adjusted according to the processing capacity of the edge nodes, whch makes our solution more versatile. In particular, as will be described in Section 4, our solution includes dynamic traffic control mechanism based on the prediction result which is not fully incorporated in [8] and [9]. Thus, our attention-based LSTM algorithm and IoT-Cloud architecture is more suitable for intelligent mobile traffic prediction and control.
4. Intelligent mobile-traffic-flow control for IoT-cloud architecture
A single edge cloud can only serve a limited number of users (i.e., a cell of a mobile network). In large cities where thousands of users or equipment are involved in many mission critical tasks, cooperation and interaction among multi-cells, multi-edge clouds, and the remote cloud can be crucially important. For instance, a rapid increase in the number of users and equipment accessing the internet can cause an unprecedented rise in mobile-traffic volume. This can easily affect the load-balancing, hence severely undermining the integrity of the entire network. More importantly, it goes against the goal of achieving high reliability and low latency as required by the uRLLC. Consequently, this has prompted us to further extend our intelligent resource allocation with mobile-traffic-flow prediction and control that goes beyond a single-site structure. Therefore, we propose a multi-site IoT-cloud extension for an intelligent mobile-traffic control as described below.
4.1. Multi-Site IoT-Cloud architecture
Fig. 4 shows our proposed multi-site cloud structure where an IoT-Cloud module uses the same concept presented earlier in Section 2.2 (i.e., an IoT-Cloud architecture based on uRLLC). In a large-scale network supporting uRLLC, each cell contains intelligent equipment of an uncertain quantity. While each equipment can communicate with the edge cloud at the edge of a cell, it can also communicate with other edge clouds. Normally, each equipment tends to select the edge-cloud offering the best communication link (usually, the nearest one) within a given time frame unless the remote cloud dispatches it in advance, based on the network congestion. In this case users will lose the accuracy rate due to packet loss when communicating with others. Often, it is called network switching.
Fig. 4.
Intelligent mobile-traffic-flow control for IoT-cloud architecture
In the proposed multi-site scheme, each edge cloud predicts the mobile-traffic flow as described in Section 3 and reports the network status to the remote cloud. The remote cloud then invokes the mobile-traffic-flow control process by intelligently dispatching and allocating the communication and computing resources across the whole network. This enables more efficient interactions between users.
4.2. Cognitive engine
We introduce a cognitive engine to implement a high-performance artificial intelligence algorithm (including the mobile-traffic-flow prediction algorithm based on our attention-based LSTM scheme as described in Section 3) with the ability to store users’ data in large quantities. With the help of Internet of Things traffic data in the edge clouds and remote clouds and our dynamic traffic-flow prediction, computing and data analysis can be offered with high precision. The cognitive engine can be divided into two types: resource cognitive engine and data cognitive engine. The detailed functions of the cognitive engine can be found in our previous study [20].
4.3. Intelligent mobile-traffic control
There are three functional elements needed to meet the uRLLC requirements. These are: mobile-traffic prediction, communication resource scheduling, and computational resource allocation. To accurately execute these functional elements of the remote cloud, the edge cloud and user devices must be well synchronized. More specifically, a user device can reduce the flow of the backbone network when operating with simple local computing and offline caching. Furthermore, by participating in multi-site communications (with the help of network switching) mobile equipment can avoid local congestion. Edge clouds can cache and forward data in a lightweight manner. They can communicate with surrounding nodes, predict the mobile-traffic flow of a single cell, and notify the remote cloud in advance as an early warning in case of high-peak traffic flow. After completing complex cognitive computing, the remote cloud predicts the mobile-traffic flow on a large scale and dynamically dispatches the resources. With the help of a cognitive engine, balancing loads and stable communications of the whole network can be accomplished and the computing results returned as a feedback.
(1). Edge-cloud selection
In order to realize an intelligent mobile-traffic control based on the traffic-prediction algorithm of Section 3, we define a simple edge-cloud selection algorithm. First, we assume that the predicted traffic value for every edge cloud in the next time slot is Trafi = {Traf1, Traf2, …, TrafE }, where E represents the total number of edge clouds. This means that the predicted traffic flow of the whole network received by the cloud is . Next, we assume sgni represents the current signal strength of a user accessing the ith edge cloud. The choice of a user to access to the ith edge cloud can be shown as:
| (7) |
This means that users will choose to send requests or offload computational tasks to the edge cloud with a lower traffic flow or a better signal strength and this would be beneficial for the load balancing of the network.
(2). Traffic-adaptive resource allocation
In order to reduce the average delay of a single edge cloud and ensure data-transmission efficiency, we define a traffic-adaptive resource-allocation mechanism within an edge cell. Thus, all the cells can carry out parallel computing and the entire network can use the above edge-cloud selection mechanism to achieve a better optimization.
In our traffic-adaptive resource-allocation algorithm, we dynamically allocate a sub-carrier power to every device based on the predicted peak traffic flow [17], as shown below:
| (8) |
To ensure fairness amongst users, we also introduce the following proportional fairness constraints:
| (9) |
where K is the total number of devices, N represents the total number of available sub-carriers, and B represents the total system bandwidth. Let’s assume the channel response of sub-carrier n for device k is hk,n. Then, the amplitude of the channel response is |hk,n| and the channel-gain matrix . N0B/N represents the noise power of each sub-carrier, ck,n is the indicative factor. Constraint C2 represents a range of sub-carrier power allocation, i.e., ck,n = 1 means that the sub-carrier n is allocated to the device k, while ck,n = 0 indicates otherwise. pk,n is the allocated power of the sub-carrier n for the device k, which is also the power resource we need to optimize. C4 represents the time-slot proportional fairness constraint where the data rate RK of device k can be expressed as:
| (10) |
When the network loss is l in time slot t, we can easily obtain the real data rate Pk of device k, according to the following formula:
| (11) |
The objective function of the above optimization model is the system and rate capacity, i.e., the predicted traffic flow Pt of the edge cloud in time slot t, as shown in formula (12).
| (12) |
Then, we can obtain the average delay of the cell. can be divided into two parts, including the transmission delay and propagation delay , as shown in formula (13) below:
| (13) |
5. Experimental results
We trained and tested our model based on the historical mobile-traffic-flow data set cached at our local server (edge cloud), in turn based on the attention-LSTM algorithm. Then, the trained prediction model was deployed at an edge-cloud to predict the traffic flow dynamically. Our experimental set up was based on a closed room containing: a cloud server, 3 local servers, and 15 smart devices. A total of 5000 pieces of data were collected from the historical mobile traffic-flow data set, involving time, packet quantity, packet size, device log, etc. The 5000 pieces of data were divided into a training set (3350 pieces) and a testing set for the prediction (1650 pieces).
Fig. 5(a) shows the root-mean-square error (RMSE) and loss results of the training and prediction data sets, which have been referred to as training RMSE and prediction RMSE in this figure. For both cases, the RMSE and data losses after initially falling considerably with increasing memory time slot, continue to drop at a slower pace up to 300 s. While the training RMSE continues to drop, the prediction RMSE begins to increase. This indicates that after a memory time slot exceeds 300 s, the prediction results become out-fitted (i.e., less accurate). Thus, we use 300 s as the best input memory time slot in order to build our attention-based LSTM model to predict traffic flow in our subsequent experiments. The comparison results consisting of the real, training, and prediction traffic flows are shown in Fig. 5(b). Although there is insufficient historical training data, it is clear that the trend of the peaks and valleys of the predicted mobile-traffic flow is almost identical to the actual one. Fig. 5(c) compares different traffic-flow prediction algorithms, including backpropagation (BP) [18], autoregressive integrated moving average (ARIMA) [19] and attention-based LSTM (with a memory time slot of 300 s for our approach). The experimental results show that our algorithm is capable of producing a far better traffic-prediction accuracy. This can consequently contribute to the effectiveness of the traffic-flow control, which will be examined next.
Fig. 5.
Comparison of the predicted traffic flow and the real traffic flow
Subsequently, we deployed our mobile-traffic-flow prediction algorithm on a cloud server, and scheduled the wireless access resources based on the predicted traffic peak [20]. In these experiments, we evaluate the packet-loss rates and average transmission latency of service requests from specific equipment received by the cloud. The results are shown in Figs. 6(a) and Fig. 6(b), respectively. Due to the small packets sent by each device and the medium size processing capacity of our server, we can only compare the results based on a single user’s requests under the following two modes of operation: 1) Intelligent Prediction and Control and 2) Traditional Transmission (without prediction and control). It is clear that the packet-loss rate in the traditional mode rises significantly as the mobile-traffic flow increases, which will reduce accuracy of computation and negatively influence QoE. We should point out that in our experiments, we obtain our packet-loss rates and average latency results using the least-squares method with two iterations. As indicated in Fig. 6, the proposed architecture can effectively improve both the packet-loss rate and the latency as compared with the traditional mode.
Fig. 6.
Performance evolution
Under the proposed intelligent mobile-traffic-flow control strategy, the cloud will initiate multi-site cooperation as soon as it receives notification of a peak mobile-traffic flow, which is sent by certain edge clouds. A specific user device will be notified to access another idle edge cloud to compute the tasks-transmit service. This greatly balances network load and guarantees stable transmission of the users’ requests as well as effectively control the latency. However, when there is low mobile-traffic flow, the latencies of both schemes are almost identical. This is mainly because users tend to select the best edge cloud (e.g., nearest) for communication when the channel is idle. As the mobile-traffic flow increases, the network transmission becomes saturated. Under these conditions, the latency of both schemes increases exponentially, but the proposed approach would be able to control transmission latency more effectively.
6. Open Issues
Although we have discussed the prediction algorithm and the control architecture of mobile-traffic flow targeting the uRLLC service, there are still many open issues that will need to be explored in the future, e.g.,
(1). Interactive map and prediction of the users’ mobility
Radio maps are an important component of 5G. They can provide useful information in the service area about radio signal strength, channel conflict, channel interference, etc. Nodes representing edge clouds are fixed, but the mobile-network environment changes easily. This requires a continuous update of the radio maps so that users can dynamically choose the best edge cloud. Obviously, users’ mobility is an important factor for predicting mobile-traffic flow, which helps the remote cloud to ascertain the peak of the group in a certain place in advance [21]. In this case, the mobile-traffic flow received by an edge cloud in densely populated areas can possibly be transferred to another edge cloud for processing. This helps to balance the network load.
(2). Edge-cloud sharing strategy
Due to the mobility of users, the popularity of content, and the high reliability of uRLLC requests, latency can be reduced to an acceptable level. Therefore, the mobile-traffic-flow prediction and control strategy realized in a single-site cell can be used repetitively and shared with other edge clouds. Learning algorithms must be deployed in each edge-cloud in order to provide a more intelligent interactions among different base-stations and to learn an appropriate resource-dispatch mode. The same edge cloud will predict the contents requested in advance or predict the mobile-traffic flow of other edge clouds.
(3). Risk perception of the Remote cloud
Users’ requests or data transmissions have priority. The events with small probability, such as alarms and notices, might involve safety concerns. Therefore, the system must anticipate risks in the process of predicting and controlling mobile-traffic flow by predicting the priority of the users’ requests [22]. Limited communication and computing resources will be allocated to each task. We can use an online learning algorithm (such as Q-Learning) in machine learning for remote clouds on the basis of predicting mobile-traffic flow. By training and learning historical risk data sets, shared wireless resources in networks can be dispatched in the future.
7. Conclusion
This paper summarizes the three application scenarios of 5G, namely eMBB, mMTC and uRLLC. We have proposed an IoT-Cloud-based architecture in support of the uRLLC application scenario. In particular, we have described the issue of mobile-traffic flows and their effect on the interaction between user devices, edge clouds and remote clouds. To achieve high reliability and low latency of communications, we first use an attention-based LSTM algorithm to predict the mobile-traffic flow in a single-site mode. Then, the remote cloud collects the traffic-flow predictions of multiple sites. With support of the cognitive engine and mobile-traffic control modules, the mobile-traffic flow for the entire network is predicted and controlled intelligently. Based on an IoT-Cloud, the performance of the proposed traffic-adaptive resource allocation algorithm is then evaluated. The experimental results verify that our mobile-traffic-flow prediction algorithm can indeed accurately predict mobile-traffic flow and is therefore capable of effectively reducing both the latency and the packet-loss rate. Finally, we discuss some open issues, including: interactive maps and prediction of the users’ mobility, edge-cloud sharing strategies, and the risk perception of remote clouds. This offers deeper research orientations in mobile-traffic prediction.
In the future, we will further verify our intelligent mobile-traffic-flow control architecture, based on an IoT-Cloud. Furthermore, the prediction of users’ mobility, which can play a major role in mobile-traffic-flow prediction, will also be investigated. The algorithm for dispatching communication resources and computing resources, together with the corresponding allocation scheme, will be further detailed.
Author Bio
Min Chen [SM’09] is a professor in School of Computer Science and Technology at Huazhong University of Science and Technology. He has more than 300 paper publications, including 200+ SCI papers, 80+ IEEE Trans./Journal papers, 32 ESI highly cited papers and 10 ESI hot papers. His Google Scholars Citations reached 15,600+ with an h-index of 62 and i10-index of 195. His top paper was cited 1800+ times. He is an IEEE Senior Member since 2009. He got IEEE Communications Society Fred W. Ellersick Prize in 2017. He was selected as Highly Cited Research at 2018 (computer science category). He is a Series Editor for IEEE Journal on Selected Areas in Communications on Network Softwarization \& Enablers. He is a technical Editor for IEEE Network, and an Editor for IEEE Transactions on Cognitive Communications and Networking. His research focuses on cognitive computing, 5G Networks, and mobile edge computing, etc.
Yiming Miao [M’18] received the B.Sc. degree in College of Computer Science and Technology from Qinghai Univerisity, Xining, China in 2016. Currently, she is a Ph.D candidate in School of Computer Science and Technology at Huazhong University of Science and Technology (HUST). Her research interests include IoT sensing, healthcare big data and emotion-aware computing, etc.
Hamid Gharavi [F’92] received his Ph.D. degree from Loughborough University, United Kingdom, in 1980. He joined the Visual Communication Research Department at AT\&T Bell Laboratories, Holmdel, New Jersey, in 1982. He was then transferred to Bell Communications Research (Bellcore) after the AT\&T-Bell divestiture, where he became a consultant on video technology and a Distinguished Member of Research Staff. In 1993, he joined Loughborough University as a professor and chair of communication engineering. Since September 1998, he has been with the National Institute of Standards and Technology (NIST), U.S. Department of Commerce, Gaithersburg, Maryland. He was a core member of Study Group XV (Specialist Group on Coding for Visual Telephony) of the International Communications Standardization Body CCITT (ITU-T). His research interests include smart grid, wireless multimedia, mobile communications and wireless systems, mobile ad hoc networks, and visual communications. He holds eight U.S. patents and has over 150 publications related to these topics. He received the Charles Babbage Premium Award from the Institute of Electronics and Radio Engineering in 1986, and the IEEE CAS Society Darlington Best Paper Award in 1989. He was the recipient of the Washington Academy of Science Distinguished Career in Science Award for 2017. He served as a Distinguished Lecturer of the IEEE Communication Society. He has been a Guest Editor for a number of special issues of the Proceedings of the IEEE, including Smart Grid, Sensor Networks & Applications, Wireless Multimedia Communications, Advanced Automobile Technologies, and Grid Resilience. He was a TPC Co-Chair for IEEE SmartGridComm in 2010 and 2012. He served as a member of the Editorial Board of Proceedings of the IEEE from January 2003 to December 2008. From January 2010 to December 2013 he served as Editor-in-Chief of IEEE Transactions on CAS for Video Technology. He is currently serving as the Editor-in-Chief of IEEE Wireless Communications.
Long Hu [M’15] received Ph.D. degrees in School of Computer Science and Technology at Huazhong University of Science and Technology (HUST). He has also received his Master and B.S. degree in HUST. He is the Publication Chair for 4th International Conference on Cloud Computing (CloudComp 2013). Currently, his research includes 5G Mobile Communication System, Big Data Mining, Marine-Ship Communication, Internet of Things, and Multimedia Transmission over Wireless Network, etc.
Iztok Humar [SM’09] received B.Sc., M.Sc. and PhD degrees in the field of Telecommunications at the Faculty of Electrical Engineering, University of Ljubljana, Slovenia, in 2000, 2003, and 2007 respectively. He also received PhD in the field of Information Management at the Faculty of Economics, University of Ljubljana, Slovenia, in 2009. Currently, he is an Associate professor at the Faculty of Electrical Engineering, where he lecturers on Design, Modeling and Management of Communication Networks on graduate and postgraduate study. He was a supervisor of many undergraduate students and some postgraduate students. His main research topics include the energy efficiency of wireless networks, cognitive computing for communications with applications and modeling of network loads and traffic for QoS/QoE. In 2010, he was a three-month Visiting Professor and Researcher at Huazhong University of Science and Technology, Wuhan, China. He is a Senior member of IEEE Communication Society from 2009 and a member of Electrical Association of Slovenia. He served as IEEE Communication Society of Slovenia Chapter Chair for 10 years.
References
- 1.Sachs J, “5G Ultra-Reliable and Low Latency Communication,” 2017. URL: http://cscn2017.ieee-cscn.org/files/2017/08/Janne_Peisa_Ericsson_CSCN2017.pdf
- 2.Ministry of information industry, “China’s telecommunications industry in the first half of 2018,” 2018. URL: http://www.199it.com/archives/751171.html
- 3.Chen M, Qian Y, Hao Y, Li Y, Song J, “Data-Driven Computing and Caching in 5G Networks: Architecture and Delay Analysis”, IEEE Wireless Communications, vol. 25, no. 1, pp.70–75, February 2018. [Google Scholar]
- 4.Wang H, Wu Y, Min G, Xu J, Tang P, “Data-driven Dynamic Resource Scheduling for Network Slicing: A Deep Reinforcement Learning Approach,” Information Sciences, vol. 498, pp. 106–116, 2019. [Google Scholar]
- 5.Abadi A, Rajabioun T, Ioannou PA, “Traffic Flow Prediction for Road Transportation Networks With Limited Traffic Data,” IEEE Transactions on Intelligent Transportation Systems, vol. 16, no. 2, pp.653–662, 2015. [Google Scholar]
- 6.Zhang G, Cao Y, Wang L, Li D, “Operation Cost Minimization for Base Stations With Heterogenous Energy Supplies and Sleep-Awake Mode: A Two-Timescale Approach,” IEEE Transactions on Cognitive Communications and Networking, vol. 4, no. 4, pp.908–918, 2018. [Google Scholar]
- 7.Liang Ying-Chang, “Editorial of IEEE Transactions on Cognitive Communications and Networking,” IEEE, vol. 5, no. 1, March 2019. DOI: 10.1109/TCCN.2019.2899276 [DOI] [Google Scholar]
- 8.Zhou Y, Fadlullah ZM, Mao B, Kato N, “A Deep-Learning-Based Radio Resource Assignment Technique for 5G Ultra Dense Networks,” IEEE Network, vol.32, no.6, pp.28–34, 2018. [Google Scholar]
- 9.Feng J, Chen X, Gao R, Zeng M, Li Y, “DeepTP: An End-to-End Neural Network for Mobile Cellular Traffic Prediction,” IEEE Network, vol.32, no.6, pp.108–115, 2018. [Google Scholar]
- 10.Greff K, Srivastava RK, Koutník J, et al. , “LSTM: A Search Space Odyssey,” IEEE Transactions on Neural Networks & Learning Systems, vol.28, no.10, pp.2222–2232, 2015. [DOI] [PubMed] [Google Scholar]
- 11.Lent R, “Resource Selection in Cognitive Networks With Spiking Neural Networks,” IEEE Transactions on Cognitive Communications and Networking, vol.4, no.4, pp.860–868, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Anand A, De Veciana G, Shakkottai S, “Joint Scheduling of URLLC and eMBB Traffic in 5G Wireless Networks,” IEEE INFOCOM 2018, pp.1970–1978, Honolulu, HI, USA, 16–19 April 2018. [Google Scholar]
- 13.Cheng X, Wu Y, Min G, Zomaya AY, “Network Function Virtualization in Dynamic Networks: A Stochastic Perspective,” IEEE Journal on Selected Areas in Communications, vol. 36, no. 10, pp. 2218–2232, 2018. [Google Scholar]
- 14.Miao W, Min G, Wu Y, Huang H, Zhao Z, Wang H, Luo C, “Stochastic Performance Analysis of Network Function Virtualisation in Future Internet,” IEEE Journal on Selected Areas in Communications, vol. 37, no. 3, pp. 613–626, 2019. [Google Scholar]
- 15.Chen M, Hao Y, Gharavi H, Leung V, “Cognitive Information Measurements: A New Perspective”, Information Science, Vol. 505, pp. 487–497, December 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Bahdanau D, Cho K, Bengio Y, “Neural Machine Translation by Jointly Learning to Align and Translate,” CoRR, vol. abs/1409.0473, 2015. [Google Scholar]
- 17.Zou T, Meng Q, Cong H, “Mobile Communication Technology and Application,” Tsinghua University Press, Beijing, China, 2013. ISBN: 9787302329251 [Google Scholar]
- 18.Wang Z, Wang B, Liu C, Wang W, “Improved BP Neural Network Algorithm to Wind Power Forecast,” The Journal of Engineering, vol.2017, no.13, pp.940–943, 2017 [Google Scholar]
- 19.Biernacki A, “Improving Quality of Adaptive Video by Traffic Prediction with (F)ARIMA Models,” Journal of Communications and Networks, vol.19, no.5, pp.521–530, 2017. [Google Scholar]
- 20.Chen M, Miao Y, Jian X, Wang X, Hu L, “Cognitive-LPWAN: Towards Intelligent Wireless Services in Hybrid Low Power Wide Area Networks”, IEEE Trans. Green Communications and Networking, Vol. 3, No. 2, pp. 407–417, June 2019. [Google Scholar]
- 21.Chen M, Hao Y, Lai C, Wu D, Li Y, Hwang K, “Opportunistic Task Scheduling over Co-Located Clouds in Mobile Environment”, IEEE Trans. Service Computing, Vol. 11, No. 3, pp. 549–561, 2018. [Google Scholar]
- 22.Ou Y, Wang X, Jiang J, Wei H, Huang C, Hsu K, “Simulator training to drive the risk perception of the reliability and validity,” IEEE International Conference on Applied System Invention (ICASI), pp.374–377, Chiba, Japan, 13–17 April 2018. [Google Scholar]







