© 2015 International Association for Mathematics and Computers in Simulation (IMACS). Published by Elsevier B.V. All rights. A novel method based on minimizing the Euclidean distance is proposed for generating a well-distributed Pareto set in multi-objective optimization for a class of ergodic controllable Markov chains. The proposed approach is based on the concept of strong Pareto policy. We consider the case where the search space is a non-strictly convex set. For solving the problem we introduce the Tikhonov's regularization method and implement the Lagrange principle. We formulate the original problem introducing linear constraints over the nonlinear problem employing the c-variable method and constraining the cost-functions allowing points in the Pareto front to have a small distance from one another. As a result, the proposed method generates an even representation of the entire Pareto surface. Then, we propose an algorithm to compute the Pareto front and provide all the details needed to implement the method in an efficient and numerically stable way. As well, we prove the main Theorems for describing the dependence of the saddle point for the regularizing parameter and analyzes its asymptotic behavior. Moreover, we analyze the step size parameter of the Lagrange principle and also its asymptotic behavior. The suggested approach is validated theoretically and verified by a numerical example related to security patrolling that present a technique for visualizing the Pareto front.