Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Stefan Bosse
University of Bremen, Dept. Mathematics & Computer Science, Bremen, Germany
sbosse@uni-bremen.de
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
This work demonstrates the benefits of agent-based simulation and learning agents for the self-organised and decentralised optimisation of traffic and logistics flows in cities
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
This work addresses three paradigms to create smart city control:
Cooperating and interacting Multi-agent Systems;
Reinforcement Learning (RL);
Self-organisation and self-adaptivity.
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Traffic and logistics flow control should be achieved on three levels:
Ensemble control by the environment
Individual control by driving entities. e.g.,
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
There are physical and computational agents handled by one unified agent model
Physical Agents. Coupled to mobile platform and representing physical entities (humans, vehicles, products, machines, ..)
Computational Agents. Representation and implementation of mobile software
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Activity-Transition Graph (ATG) behaviour and data model of an agent for a specific class AC (left). Physical and computational agents differ in their action set (middle). Physical and computational agents can communicate with each other (right)
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
The proposed hybrid parameterised agent architecture combining reactive state-based action selection with RL
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
percept:Sen×Per→Pernext:St×Per×Par×R×C→Staction:St×Par→Act1rl:r×Per×→Act2reward:Act2×Per×Par→r[−1,1]fusion:Act1×Act2→Act
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
The utility function u(S): S → r provides the necessary input for the reinforcement learning instance (reward function)
The utility function uses a set of internal and external state variables S derived from sensor input (Per):
type states S = { v0: Normalised average speed, ds0: Distances to s={front, back, left, right} neighbour vehicles, de, Δde: Distance to destination and delta change (progress), td: Direction to destination (0-360 degree), r0: Direction of vehicle (0-360 degree), qt0: Queuing time, dd: Allowed driving and turning directions, P: Set of possible paths from current position to destination}
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Vehicle Agent → Rule-based short-range navigation
Navigation Agent → Reinforcement Learning of long-range navigation for path length and time optimisation
Traffic Control Agent → Flow sensing and control (traffic lights and signs)
Environmental sensors used by agents and agent interaction
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Act = [ Moving one step left, right, backward, or ahead: |Δ|=1, Satisfy distance constraints: Increasing or decreasing the vehicle speed, Follow short-range Δ displacement vector (minΔ), Stopping movement]
Act = [ Change direction to N/S/W/E, Keep direction (forced), Change vehicle speed, Change destination, Escape blocking situations]
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Although the learning and adaptive long-range traffic routing can be deployed in real world, simulation is used to investigate training and impact of the proposed approach
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Core component: The JavaScript Agent Machine is a portable and powerful agent processing platform written in JavaScript and capable to execute JavaScript agents
SEJAM extends the agent platform JAM with a simulation and visualisation layer
SEJAM support the concept of closed-loop simulation for augmented virtuality
Mobile and non-mobile devices executing the JAM platform can be connected with the virtual simulation world (via the Internet)
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Virtual simulation world (Simulator SEJAM) coupled to real worlds via the Internet and unified agent processing platforms (JAM)
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
The simulation model and the agents are programmed entirely in JavaScript
It consists of
Physical and computational agents can be modelled the same way!
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
// Agent Class Constructorsfunction world (options) { this.XX=xx this.act={} this.trans={} this.on={} this.next=init }function vehicle (options) { this.XX=xx this.act={} this.trans={} this.on={} this.next=init }function navigator (options) { this.XX=xx this.act={} this.trans={} this.on={} this.next=init }// Simulation World Model Descriptormodel = { agents : { world : { behaviour:world, visual : { .. }, .. }, resources : { street : { .. }, place : { .. }, signal : { .. }, .. }, nodes : { world : { .. }, vehicle : { .. } , navigator : { .. }, parameter : { .. }, world : { init : {..}, map : {..}, resources : {..}, patchgrid : {..} } }
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Path eff.: (Left) Without pre-trained agents (Right) With pre-trained agents
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Progress of global reward without pre-trained agents
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
In contrast to common traffic management controlling traffic lights and signals only, this work addressed traffic flow optimisation on micro-level by adapting decision making processes of vehicles, primarily long-range navigation and re-routing, optionally with vehicle speed control.
Training of reinforcement learning navigation agents by thousands of trial-and-error cycles requires a long time to reach a satisfying navigation strategy better than random walk and is only possible in simulation worlds. Otherwise domestic traffic would collapse if performed in real world.
Simulation results from an agent-based simulation of an artificial urban area show that the deployment of such a micro-level vehicle control just by individual decision making, learning, and re-routing based on local environmental sensors can reach near optimal routing still under high traffic densities (regarding total route length and travelling times).
Stefan Bosse: Self-adaptive Traffic and Logistics Flow Control using Learning Agents and Ubiquitous Sensors
Thank you for your attention. All questions are welcome!