Coordination Between Traffic Signals Based On Cooperative: S.S. Shamshirband, H. Shirgahi, M.Gholami and B. Kia
Coordination Between Traffic Signals Based On Cooperative: S.S. Shamshirband, H. Shirgahi, M.Gholami and B. Kia
ISSN 1818-4952
© IDOSI Publications, 2008
1
Department of Computer, Islamic Azad University, Chalous Branch, Chalous, Iran
2
Department of Computer, Islamic Azad University, Jouybar Branch, Jouybar, Iran
3
Iranian Academic Center for Education, Culture and Research Mazandaran Branch, Sari, Iran
4
Department of Science, Islamic Azad University, Chalous Branch, Chalous, Iran
Abstract: The single traffic signal control agent improves its control ability with the Multiagents-learning
method. This paper proposes a new cooperative learning method; called weighted strategy sharing (WSS) is
presented. In this method, each agent measures the expertness of its teammates and assigns a weight to
their knowledge and learns from them accordingly. The presented methods are tested on three traffic lights.
Also, the effect of the communication noise, as a source of uncertainty, on the cooperative learning method
is studied. Moreover, the Qtable of one of the cooperative agents is changed randomly and its effects on
the presented methods are examined. Results using cooperative traffic agents are compared to results of
control simulations where non-cooperative agents were deployed. The result indicates that the new
coordination method proposed in this paper is effective.
Q̂ = f((AR,QN,QC),a, θ ) (1)
The decision making process based on a set of Where, n is the number of the actions; Q (a) is the
fuzzy rules which takes into account the traffic evaluation value of action a; τ is a positive number
Conditions with the current and next phases. The named as temperature. The higher the temperature, the
general format of the fuzzy rules is as follows: more average every action is selected.
ej
Expertness criteria: In the WSS method, Wij is a Wij = n
(5)
measure of agent reliance on the knowledge and the ∑ ek
k =1
experiences of agent. Here we argue that this weight is a
function of the agents’ relative expertness. In the
where n is the number of the agents and ek is the
strategy sharing method, expertness of the agents are
amount of the expertness of agent k. In this method,
assumed to be equal. Some studies used the user
effects of agent j knowledge on all learners are equal, i.e.
judgment for specifying the expert agent. This method
requires continuous human supervision.However, some
W 1j = W 2j =... W nj
studies specified the expert agents by means of their
successes and failures during current moves and Also all of Q-tables become homogeneous after
considered the expertness criterion as an algebraic sum each cooperation step.
528
World Appl. Sci. J., 5 (5): 525-530, 2008
IMPLEMENTATION 40
35
We have constructed a prototype traffic simulator
30
program to test the efficiency of the coordination
5
10
15
20
25
30
35
40
45
50
intersections, traffic light control agent and vehicles. cycle
CONCLUSION
529
World Appl. Sci. J., 5 (5): 525-530, 2008
2. Sanchez, S.M. and T.W. Lucas, 1972. Exploring the 10. Sutton, R.S., 1988. Learning to predict by the
world of agent based simulations: simple models, methods of temporal differences. Mach. Learn,
complex analyses, Proceedings of the Simulation 3: 9-44.
Conference, pp.116-126. 11. Sutton, R.S., 1998. Machine Learning: Special Issue
3. Zhongzhi, S., 1998. The Advanced Artificial on Reinforcement Learning. Cambridge, MA: MIT
Intelligence. Beijing: Science Press, Chapter 10, Press, 8: 3-4.
pp: 223-226. 12. Yamaguchi, T., Y. Tanaka and M. Yachida, 1997.
4. Pan, G.C. and B. Maddox, 1995. A Framework for Speed up reinforcement learning between two
Distributed Reinforcement Learning, Lecture Notes agents with adaptive mimetism. Proc. IEEE Conf.
in Artificial Intelligence, Adaptation and Learning in Intl. Robot. Syst. (IROS), pp: 594-600.
Multi-Agent Systems, pp: 97-112.
13. Friedrich, H., M. Kaiser, O. Ragalla and R. Dillmann,
5. Liu, Z., 2007. A Survey of Intelligence Methods in
1996. Learning and communication in multi-agent
Urban Traffic Signal Control, International Journal
systems. Distributed Artificial Intelligence Meets
of Computer Science and Network Security, 7(7):
105-112. Machine Learning. Weiss, G. (Ed.). New York:
6. Lucia, A. and C. Bazan, 1994. Traffic Signal Springer-Verlag, 1221: 259-275.
Coordination Based on Distributed Problem 14. Watkins, C.J., 1989. Learning from delayed rewards.
Solving. 7’ Fanfares Symposium on Transportation Ph.D. dissertation, King’s College, Cambridge, UK.
Systems: Theory and Application of Advanced 15. Watkins, C.J. and P. Dayan, 1998. Q-learning
Technology, Tianjin, China, pp: 957-962. (technical note). Machine Learning: Special Issue
7. Ming, L.X. and F.Y. Wang, 2001. Study of city area on Reinforcement Learning. Cambridge, MA: MIT
traffic coordination control on the basis of agent. Press, pp: 55-68.
F’roceedings of the IEEE intelligent Transportation 16. Tan, M., 1993. Multi-agent reinforcement learning:
Systems Conference, Singapore, September, Independent vs. cooperative agents. In: Proc. 10th
pp: 758-761. Intl. Conf. Machine Learning, Amherst, MA.
8. Ossowaki, S., J. Cuena and A. Garcia, 1998. A Case 17. Friedrich, H., 1996. Learning and communication in
of Multiagent Decision Support: Using multi-agent systems, in Distributed Artificial
autonomous agents for urban traffic control. Intelligence Meets Machine Learning, Weiss, G.,
Lecture Notes in Artificial Intelligence, 1484: (Ed.). New York: Springer-Verlag, 1221: 259-275.
100-111.
9. Chen, W. and K. Decker, 2004. Developing
Altcmative Mechanisms for Multiagent
Coordination. Lecture Notes in Computer Science,
Springer-Verlag Heidelberg, 2413: 63-76.
530