TY - JOUR AU1 - Xu, Yanggang AU2 - Hong, Weijie AU3 - Zha, Jirong AU4 - Chen, Geng AU5 - Zheng, Jianfeng AU6 - Hsia, Chen-Chun AU7 - Chen, Xinlei AB - Abstract:In disaster scenarios, establishing robust emergency communication networks is critical, and unmanned aerial vehicles (UAVs) offer a promising solution to rapidly restore connectivity. However, organizing UAVs to form multi-hop networks in large-scale dynamic environments presents significant challenges, including limitations in algorithmic scalability and the vast exploration space required for coordinated decision-making. To address these issues, we propose MRLMN, a novel framework that integrates multi-agent reinforcement learning (MARL) and large language models (LLMs) to jointly optimize UAV agents toward achieving optimal networking performance. The framework incorporates a grouping strategy with reward decomposition to enhance algorithmic scalability and balance decision-making across UAVs. In addition, behavioral constraints are applied to selected key UAVs to improve the robustness of the network. Furthermore, the framework integrates LLM agents, leveraging knowledge distillation to transfer their high-level decision-making capabilities to MARL agents. This enhances both the efficiency of exploration and the overall training process. In the distillation module, a Hungarian algorithm-based matching scheme is applied to align the decision outputs of the LLM and MARL agents and define the distillation loss. Extensive simulation results validate the effectiveness of our approach, demonstrating significant improvements in network performance, including enhanced coverage and communication quality. TI - Scalable UAV Multi-Hop Networking via Multi-Agent Reinforcement Learning with Large Language Models JF - Computing Research Repository DO - 10.48550/arxiv.2505.08448 DA - 2025-05-13 UR - https://www.deepdyve.com/lp/arxiv-cornell-university/scalable-uav-multi-hop-networking-via-multi-agent-reinforcement-WHBG07WbW4 VL - 2025 IS - 2505 DP - DeepDyve ER -