TY - JOUR
T1 - Multiobjective Markov chains optimization problem with strong Pareto frontier
T2 - Principles of decision making
AU - Clempner, Julio B.
AU - Poznyak, Alexander S.
N1 - Publisher Copyright:
© 2016 Elsevier Ltd
PY - 2017/2/1
Y1 - 2017/2/1
N2 - In this paper, we present a novel approach for computing the Pareto frontier in Multi-Objective Markov Chains Problems (MOMCPs) that integrates a regularized penalty method for poly-linear functions. In addition, we present a method that make the Pareto frontier more useful as decision support system: it selects the ideal multi-objective option given certain bounds. We restrict our problem to a class of finite, ergodic and controllable Markov chains. The regularized penalty approach is based on the Tikhonov's regularization method and it employs a projection-gradient approach to find the strong Pareto policies along the Pareto frontier. Different from previous regularized methods, where the regularizator parameter needs to be large enough and modify (some times significantly) the initial functional, our approach balanced the value of the functional using a penalization term (μ) and the regularizator parameter (δ) at the same time improving the computation of the strong Pareto policies. The idea is to optimize the parameters μ and δ such that the functional conserves the original shape. We set the initial value and then decrease it until each policy approximate to the strong Pareto policy. In this sense, we define exactly how the parameters μ and δ tend to zero and we prove the convergence of the gradient regularized penalty algorithm. On the other hand, our policy-gradient multi-objective algorithms exploit a gradient-based approach so that the corresponding image in the objective space gets a Pareto frontier of just strong Pareto policies. We experimentally validate the method presenting a numerical example of a real alternative solution of the vehicle routing planning problem to increase security in transportation of cash and valuables. The decision-making process explored in this work correspond to the most frequent computational intelligent models applied in practice within the Artificial Intelligence research area.
AB - In this paper, we present a novel approach for computing the Pareto frontier in Multi-Objective Markov Chains Problems (MOMCPs) that integrates a regularized penalty method for poly-linear functions. In addition, we present a method that make the Pareto frontier more useful as decision support system: it selects the ideal multi-objective option given certain bounds. We restrict our problem to a class of finite, ergodic and controllable Markov chains. The regularized penalty approach is based on the Tikhonov's regularization method and it employs a projection-gradient approach to find the strong Pareto policies along the Pareto frontier. Different from previous regularized methods, where the regularizator parameter needs to be large enough and modify (some times significantly) the initial functional, our approach balanced the value of the functional using a penalization term (μ) and the regularizator parameter (δ) at the same time improving the computation of the strong Pareto policies. The idea is to optimize the parameters μ and δ such that the functional conserves the original shape. We set the initial value and then decrease it until each policy approximate to the strong Pareto policy. In this sense, we define exactly how the parameters μ and δ tend to zero and we prove the convergence of the gradient regularized penalty algorithm. On the other hand, our policy-gradient multi-objective algorithms exploit a gradient-based approach so that the corresponding image in the objective space gets a Pareto frontier of just strong Pareto policies. We experimentally validate the method presenting a numerical example of a real alternative solution of the vehicle routing planning problem to increase security in transportation of cash and valuables. The decision-making process explored in this work correspond to the most frequent computational intelligent models applied in practice within the Artificial Intelligence research area.
KW - Decision making
KW - Markov chains
KW - Multi-objective
KW - Pareto
KW - Tikhonov's regularization
UR - http://www.scopus.com/inward/record.url?scp=84992065891&partnerID=8YFLogxK
U2 - 10.1016/j.eswa.2016.10.027
DO - 10.1016/j.eswa.2016.10.027
M3 - Artículo
SN - 0957-4174
VL - 68
SP - 123
EP - 135
JO - Expert Systems with Applications
JF - Expert Systems with Applications
ER -