Skip to main content
Kent Academic Repository

Deep Learning Applications in Beyond 5G Mobile Networks

Laurinavicius, Ignas (2023) Deep Learning Applications in Beyond 5G Mobile Networks. Doctor of Philosophy (PhD) thesis, University of Kent,. (doi:10.22024/UniKent/01.02.104063) (Access to this publication is currently restricted. You may be able to access a copy if URLs are provided) (KAR id:104063)

PDF
Language: English

Restricted to Repository staff only until October 2026.

Contact us about this Publication
[thumbnail of 44PhD_Thesis-17.pdf]
Official URL:
https://doi.org/10.22024/UniKent/01.02.104063

Abstract

The thesis is motivated by the evolution of mobile networks needing to match novel applications. Cloud radio access network (C-RAN) is a network architecture that solves power consumption issues, improves scaling, and reduces the operational expenditures for operators. Due to capacity limitations of fronthaul links, C-RAN fails to meet the expectations placed on it in solving the network objectives. Fog radio access network (F-RAN) has been proposed to solve the fronthaul loading. By placing radio processing and caching functionality at the edge nodes latency can be reduced and fronthaul experiences less load, increasing performance gains of the cloud infrastructure.

F-RAN faces multiple issues since the fog access points (FAPs) are a novel addition to the network architecture. FAPs have to communicate with each other to solve optimisation problems and provide a satisfactory quality of service (QoS). Further, compared to the baseband unit (BBU), FAPs are equipped with much smaller caches, so algorithms have to proactively cache content that users would like to access. Due to the stringent low latency requirement set for the future generation of mobile networks, some information cannot be exchanged, thus FAPs have to make assumptions about the global state of the environment. For this reason, the thesis looks at machine learning options, which are known to be able to predict and generalise outcomes based on historical data.

Reinforcement learning (RL) is chosen as it learns from an environment that follows a Markov decision process (MDP). Double deep Q-network (DDQN) is known to be one of the best performers, as it provides convergence guarantees for MDP environments and improved learning stability through the use of experience replay. The thesis uses C-RAN architecture as an environment for a DDQN deployment to control flexible ethernet (FlexE) calendar slot allocation as well as remote unit (RU) load-balancing. The operation of the method is assessed and its performance is compared to auto regressive integrated moving average (ARIMA) algorithm, which uses statistical network information to perform resource allocation for FlexE links.

Assuming information about the global state is unknown, the environment becomes partially observable. This is expected to be the case for F-RAN considering the latency constraints. In a partially observable Markov decision process (POMDP) this leads to uncertainty in making decisions, as the outcomes are perceived as stochastic by the agent. With the goal of maximising the number of users served by multiple FAPs in a POMDP under stringent constraints, DDQN fails to converge to a solution. For this problem, a multi-agent double deep Q-network (MA-DDQN) solution is proposed, in which FAPs exchange information about their experiences after every learning step. The method is tested in a simulated F-RAN environment and its performance is discussed.

While focusing on allocating resources provides satisfactory performance it has been shown to only be a half of the complete solution. Since FAPs are capable of caching data or applications close to the user, controlling what is being cached is equally as important. This makes the problem a capacitated facility location problem (CFLP). A method is proposed that chooses which services to cache in the fog layer in the first step and how to distribute the content to the users in the second step. An algorithm based on MA-DDQN, termed focused batch DDQN (FB-DDQN) is used to solve the optimisation problem of improving the cache hit ratio in the fog layer. FB-DDQN focuses on areas of the state space that the agents know they are uncertain with and requests other agents to share their solutions to similar states, which improves the generalisation of the less knowledgeable agent. The results are compared to the base MA-DDQN algorithm, as well as a RL algorithm that is written for this task.

Item Type: Thesis (Doctor of Philosophy (PhD))
Thesis advisor: Zhu, Huiling
Thesis advisor: Wang, Jiangzhou
DOI/Identification number: 10.22024/UniKent/01.02.104063
Uncontrolled keywords: Fog-RAN, Cloud-RAN, 6G, Deep Reinforcement Learning, Optimisation, POMDPs
Subjects: T Technology
Divisions: Divisions > Division of Computing, Engineering and Mathematical Sciences > School of Engineering and Digital Arts
Funders: University of Kent (https://ror.org/00xkeyj56)
SWORD Depositor: System Moodle
Depositing User: System Moodle
Date Deposited: 27 Nov 2023 08:34 UTC
Last Modified: 28 Nov 2023 12:35 UTC
Resource URI: https://kar.kent.ac.uk/id/eprint/104063 (The current URI for this page, for reference purposes)

University of Kent Author Information

Laurinavicius, Ignas.

Creator's ORCID:
CReDIT Contributor Roles:
  • Depositors only (login required):

Total unique views for this document in KAR since July 2020. For more details click on the image.