Effectiveness of Q-learning as a tool for calibrating agent-based supply network models

Authored by Y. Zhang, S. Bhattacharyya

Date Published: 2007

DOI: 10.1080/17517570701275390

Sponsors: No sponsors listed

Platforms: Swarm

Model Documentation: Other Narrative Flow charts Mathematical description

Model Code URLs: Model code not found

Abstract

This paper examines effectiveness of Q-learning as a tool for specifying agent attributes and behaviours in agent-based supply network models. Agent-based modelling (ABM) has been increasingly employed to study supply chain and supply network problems. A challenging task in building agent-based supply network models is to properly specify agent attributes and behaviours. Machine learning techniques, such as Q-learning, can be a useful tool for this purpose. Q-learning is a reinforcement learning technique that has been shown to be an effective adaptation and searching mechanism in distributed settings. In this study, Q-learning is employed by supply network agents to search for `optimal' values for a parameter in their operating policies simultaneously and independently. Methods are designed to identify the `optimal' parameter values against which effectiveness of the learning is evaluated. Robustness of the learning's effectiveness is also examined through consideration of different model settings and scenarios. Results show that Q-learning is very effective in finding the `optimal' parameter values in all model settings and scenarios considered.
Tags
Agent-based modelling computational techniques Distributed artificial intelligence (DAI) Q-learning supply chain management (SCM)