|
[1]胡大伟,张世鹏,刘慧甜,等.应急响应初期“卡车-无人机”联合配送路径问题[J].长安大学学报(自然科学版), 2024, 44(1): 105-119. [HU D W, ZHANG, S P, LIU
H T, et al. Routing problem of truck-drones joint
distribution in the initial stage of emergency response[J].
Journal of Chang'an University (Natural Science Edition),
2024, 44(1): 105-119.]
[2]康柳江,李浩,孙会君,等.复杂山区工程建设物资运输无人机巡航模型构建与实证研究[J].交通运输系统工程与信息,2023, 23(3): 290-299. [KANG L J, LI H,
SUN H J, et al. UAV cruising for material transportation
under engineering construction in complex mountainous
areas: Modeling and case study[J]. Journal of
Transportation Systems Engineering and Information
Technology, 2023, 23(3): 290-299.]
[3]
许云鹏,谢雅琪,于然,等.感-通-物多目标融合应急无人机路径规划方法[J]. 通信学报,2024, 45(4): 1-12.
[XU Y P, XIE Y Q, YU R, et al. Integrated perception
communication-logistics multi-objective oriented path
planning
for
emergency UAVs[J]. Journal on
Communications, 2024, 45(4): 1-12.]
[4]张洪海,李翰,刘皞,等.城市区域物流无人机路径规划[J]. 交通运输系统工程与信息,2020, 20(6): 22-29.
[ZHANG H H, LI H, LIU H, et al. Path planning for
logistics unmanned aerial vehicle in urban area[J].
Journal of Transportation Systems Engineering and
Information Technology, 2020, 20(6): 22-29.]
[5]胡小兵,杨常澍,周隽.复杂城市环境下无人机路网模型研究[J]. 交通运输系统工程与信息, 2023, 23(4):
251-261. [HU X B, YANG C S, ZHOU J. Route network
modeling for unmanned aerial vehicle in complex urban
environment[J]. Journal of Transportation Systems
Engineering and Information Technology, 2023, 23(4):
251-261.]
[6]梁吉,王立松,黄昱洲,等.基于深度强化学习的四旋翼无人机自主控制方法[J].计算机科学,2023,50(S2):
13-19. [LIANG J, WANG L S, HUANG Y Z, et al.
Autonomous control algorithm for quadrotor based on
deep reinforcement learning[J]. Computer Science, 2023,
50(S2): 13-19.]
[7]YU X, FAN Y, XUS, et al. A self-adaptive SAC-PID
control approach based on reinforcement learning for
mobile robots[J]. International Journal of Robust and
Nonlinear Control, 2022, 32(18): 9625-9643.
[8]WILLIAMS G, WAGENER N, GOLDFAIN B, et al.
Information
theoretic
MPC for model-based
reinforcement learning[C]//2017 IEEE International
Conference on Robotics and Automation (ICRA), IEEE,
2017: 1714-1721.
[9]
王硕, 李洋, 赵蕴龙, 等. 无人机航迹规划算法综述[J/OL]. 哈 尔 滨 工程大学学报, (2025-06-16)
[2025- 07- 07].
https://link.cnki.net/urlid/23.1390.
U.20250616.1544.003. [WANG S, LI Y, ZHAO Y L,
et al. A review of UAV trajectory planning algorithms[J/
OL]. Journal of Harbin Engineering University, (2025
06-16) [2025-07-07]. http://kns.cnki.net/kcms/detail/
23.1390.U.20250616.1544.003.html.]
[10] 吕超, 李慕宸,欧家骏.基于分层深度强化学习的无人机混合路径规划[J].北京航空航天大学学报,2025,51
(10): 3451-3459. [LV C, LI M C, OU J J. UAV hybrid
path planning based on hierarchical deep reinforcement
learning[J]. Journal of Beijing University of Aeronautics
and Astronautics, 2025, 51(10): 3451-3459.]
[11] 杜江涛,于家明,齐辉.无人机集群不完全信息路径规划方法[J]. 哈尔滨工程大学学报,2024,45(11): 2210
2217. [DU J T, YU J M, QI H. Incomplete information
path planning method for an UAV cluster[J]. Journal of
Harbin Engineering University, 2024, 45(11): 2210
2217.]
[12] 滕菲, 王迎春,姚永辉,等.基于深度强化学习的无人机动态避障规划[J/OL]. 北京航空航天大学学报,
(2025-05-23)[2025-07-07]. https://doi.org/10.13700/j.
bh.1001-5965.2025.0084. [TENG F, WANG Y C, YAO
Y H, et al. Dynamic obstacle avoidance planning for
UAV based on deep reinforcement learning[J/OL].
Journal of Beijing University of Aeronautics and
Astronautics, (2025-05-23) [2025-07-07]. https://doi.
org/10.13700/j.bh.1001-5965.2025.0084.]
[13] FENG Z, HUANG M, WU D, et al. Multi-agent
reinforcement learning with policy clipping and average
evaluation for UAV-assisted communication Markov
game[J]. IEEE Transactions on Intelligent Transportation
Systems, 2023, 24(12): 14281-14293.
[14] AL-HILO A, SAMIR M, ASSI C, et al. UAV-assisted
content delivery in intelligent transportation systems
joint trajectory planning and cache management[J]. IEEE
Transactions on Intelligent Transportation Systems,
2020, 22(8): 5155-5167.
[15] XI M, DAI H, HE J, et al. A lightweight reinforcement
learning-based real-time path-planning method for
unmanned aerial vehicles[J]. IEEE Internet of Things
Journal, 2024, 11(12): 21061-21071.
[16] HAARNOJA T, ZHOU A, HARTIKANEN K, et al. Soft
actor-critic algorithms and applications[J]. arXiv Preprint
arXiv, 2018: 1812.05905.
[17] SCHAUL T, QUAN J, ANTONOGLOU I, et al. Prioritized
experience replay[J]. arXiv Preprint arXiv, 2015:
1511.05952.
[18] GOK M. Dynamic path planning via Dueling Double
Deep Q-Network (D3QN) with prioritized experience
replay[J]. Applied Soft Computing, 2024, 158: 111503.
[19] ZHOU Y, YANG J, GUO Z, et al. An indoor blind area
oriented autonomous robotic path planning approach
using deep reinforcement learning[J]. Expert Systems
with Applications, 2024, 254: 124277.
[20] HASSANI H, NIKAN S, SHAMI A. Traffic navigation via
reinforcement learning with episodic-guided prioritized
experience replay[J]. Engineering Applications of
Artificial Intelligence, 2024, 137: 109147.
[21] BEARD R W, MCLAIN T W. Small unmanned aircraft:
Theory and practice[M]. Princeton: Princeton university
press, 2012.
[22] SUTTON R S, BARTO A G. Reinforcement learning: An
introduction[M]. Cambridge: MIT Press, 1998.
[23] BERROCAL E, SIERRA B, HERRERO H. Evaluating
PyBullet and Isaac Sim in the Scope of Robotics and
Reinforcement Learning[J]. 2024 7th Iberian Robotics
Conference
(ROBOT), 2024,
DOI:
10.1109/ robot61475.2024.10797383.
[24] RAFFIN A, HILL A, GLEAVE A, et al. Stable
baselines3:
Reliable
reinforcement
learning
implementations[J]. Journal of Machine Learning
Research, 2021, 22(268): 1-8.
|