diff --git a/_data/citations.yaml b/_data/citations.yaml index 83ba049..9d141d4 100644 --- a/_data/citations.yaml +++ b/_data/citations.yaml @@ -14,7 +14,7 @@ - "J\xE1nos M\xE1rk Szalai-Gindl" publisher: arXiv date: '2023-11-14' - link: https://doi.org/gs9w7h + link: https://doi.org/10.48550/arxiv.2311.08118 type: paper image: images/light.jpg description: Explainability in Graph Neural Networks (GNNs) is a new field growing @@ -45,7 +45,7 @@ - Sharath M Shankaranarayana publisher: arXiv date: '2023-10-30' - link: https://doi.org/gs9w7j + link: https://doi.org/10.48550/arxiv.2310.19573 type: paper image: images/night.jpg description: Supervised machine learning relies on the availability of good labelled @@ -72,7 +72,7 @@ - "Eugenio F. S\xE1nchez-\xDAbeda" publisher: arXiv date: '2023-09-22' - link: https://doi.org/gtrktq + link: https://doi.org/10.48550/arxiv.2309.12913 type: paper image: images/space.jpg description: ESaliency maps have become one of the most widely used interpretability @@ -105,9 +105,9 @@ - Jose Outes-Carnero - Yak Ng-Molina - Juan Ramiro-Moreno - publisher: arXiv + publisher: IAENG-IJCS date: '2023-05-24' - link: https://doi.org/gt3cgs + link: https://doi.org/10.48550/arxiv.2302.12899 type: paper image: images/cloud_city.jpg description: This paper presents a method for optimizing wireless networks by adjusting diff --git a/_data/sources.yaml b/_data/sources.yaml index 54bc70a..684ff5f 100644 --- a/_data/sources.yaml +++ b/_data/sources.yaml @@ -71,6 +71,7 @@ - Yak Ng-Molina - Juan Ramiro-Moreno description: This paper presents a method for optimizing wireless networks by adjusting cell parameters that affect both the performance of the cell being optimized and the surrounding cells. The method uses multiple reinforcement learning agents that share a common policy and take into account information from neighboring cells to determine the state and reward. In order to avoid impairing network performance during the initial stages of learning, agents are pre-trained in an earlier phase of offline learning. During this phase, an initial policy is obtained using feedback from a static network simulator and considering a wide variety of scenarios. Finally, agents can intelligently tune the cell parameters of a test network by suggesting small incremental changes, slowly guiding the network toward an optimal configuration. The agents propose optimal changes using the experience gained with the simulator in the pre-training phase, but they can also continue to learn from current network readings after each change. The results show how the proposed approach significantly improves the performance gains already provided by expert system-based methods when applied to remote antenna tilt optimization. The significant gains of this approach have truly been observed when compared with a similar method in which the state and reward do not incorporate information from neighboring cells. + publisher: IAENG-IJCS buttons: - type: paper text: Manuscript