• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Zhao, Qing (Zhao, Qing.) | Li, Zhen (Li, Zhen.) | Li, Jianqiang (Li, Jianqiang.) | Guo, Jianxiong (Guo, Jianxiong.) | Ding, Xingjian (Ding, Xingjian.) | Li, Deying (Li, Deying.)

Indexed by:

SCIE

Abstract:

It's a promising way to use Unmanned Aerial Vehicles (UAVs) as mobile base stations to collect data from sensor nodes, especially for large-scale wireless sensor networks. There are a lot of works that focus on improving the freshness of the collected data or the data collection efficiency by scheduling UAVs. Given that sensing data in certain applications is time-sensitive, with its value diminishing as time progresses based on Timeliness of Information (ToI), this paper delves into the UAV Trajectory optimization problem for Maximizing the ToI-based data utility (TMT). We give the formal definition of the problem and prove its NP-Hardness. To solve the TMT problem, we propose a deep reinforcement learning-based algorithm that combines the Action Rejection Mechanism and the Deep Q-Network with Priority Experience Replay (ARM-PER-DQN). Where the action rejection mechanism could reduce the action space and PER helps improve the utilization of experiences with high value, thus increasing the training efficiency. To avoid the unbalanced data collection problem, we also investigate a variant problem of TMT (named V-TMT), i.e., each sensor node can be visited by the UAV at most once. We prove that the V-TMT problem is also NP-Hard, and propose a 2-approximation algorithm as the baseline of the ARM-PER-DQN algorithm. We conduct extensive simulations for the two problems to validate the performance of our designs, and the results show that our ARM-PER-DQN algorithm outperforms other baselines, especially in the V-TMT problem, the ARM-PER-DQN algorithm always outperforms the proposed 2-approximation algorithm, which suggests the effectiveness of our algorithm.

Keyword:

Trajectory optimization Approximation algorithm Deep reinforcement learning Data collection

Author Community:

  • [ 1 ] [Zhao, Qing]Beijing Univ Technol, Coll Comp Sci, Beijing, Peoples R China
  • [ 2 ] [Li, Zhen]Beijing Univ Technol, Coll Comp Sci, Beijing, Peoples R China
  • [ 3 ] [Li, Jianqiang]Beijing Univ Technol, Coll Comp Sci, Beijing, Peoples R China
  • [ 4 ] [Ding, Xingjian]Beijing Univ Technol, Coll Comp Sci, Beijing, Peoples R China
  • [ 5 ] [Guo, Jianxiong]Beijing Normal Univ, Adv Inst Nat Sci, Zhuhai, Peoples R China
  • [ 6 ] [Guo, Jianxiong]BNU HKBU United Int Coll, Guangdong Key Lab AI & Multimodal Data Proc, Zhuhai, Peoples R China
  • [ 7 ] [Li, Deying]Renmin Univ China, Sch Informat, Beijing, Peoples R China

Reprint Author's Address:

  • [Ding, Xingjian]Beijing Univ Technol, Coll Comp Sci, Beijing, Peoples R China

Show more details

Related Keywords:

Source :

JOURNAL OF COMBINATORIAL OPTIMIZATION

ISSN: 1382-6905

Year: 2025

Issue: 3

Volume: 49

1 . 0 0 0

JCR@2022

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 12

Affiliated Colleges:

Online/Total:1428/10641816
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.