This Project is a traffic control system based on DQN (arxiv:1312.5602) on Vissim. It's an original implement that intelligent traffic signal control via deep reinforment learning on partial urban traffic net. Choosing fine hyper-parameters, agent could learn to how to improve the performance of global net in a long term.
VisEnv.py wrapped the orignal api into the open.ai style. For now, speed, travel time, queued vehicles count interfaces are provided. Use this like:
fron vis_env import *
env = VisEnv()
...
for epi in range(episodes):
env.reset()
env.test = True
for _ in range(steps):
next_state, reward, done = env.step(action)
env.write_summary(epi, dir)
The performance of DQN is not so good among the series of reinforcement learning algorithm, but agent are still capble to act appropriately in our traffic enviroment.
More reinforment learning models like dueling-DQN, DDPG, to further improve the performance of agent, and to solve the large discrete actions space problem. Intelligent Traffic Signal Control
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。