Menu
Publications
2026
2025
2024
2023
2022
2021
2020
2019
2018
2017
2016
2015
2014
2013
2012
2011
2010
2009
2008
2007
2006
2005
2004
2003
2002
2001
Editor-in-Chief
Nikiforov
Vladimir O.
D.Sc., Prof.
Partners
doi: 10.17586/2226-1494-2026-26-1-85-93
Clustering of the approximated Pareto front
Read the full article
Article in Russian
For citation:
Abstract
For citation:
Yurtaev A.G. Clustering of the approximated Pareto front. Scientific and Technical Journal of Information Technologies, Mechanics and Optics, 2026, vol. 26, no. 1, pp. 85–93 (in Russian). doi: 10.17586/2226-1494-2026-26-1-85-93
Abstract
In contemporary engineering and scientific practice, multi-objective optimization often facilitates the search for compromise solutions without prescribing weight coefficients or bounds, forming a Pareto front via heuristic approximation based on genetic algorithms. However, even an approximated Pareto front consists of a large set of points, which complicates analysis and selection of solutions. To organize and structure the obtained results, clustering can be employed to identify representative groups of trade-offs. The scientific novelty of the proposed clustering method lies in the combination of Ordering Points to Identify the Clustering Structure and k-means algorithms with the introduction of medoids identification, which ensures automatic noise removal and a compact representation of representative strategies. A two-stage clustering approach is proposed. At the first stage, Ordering Points to Identify the Clustering Structure algorithm is used to construct an ordered density profile and to automatically filter out noise points based on the reachability threshold. At the second stage, the k-means algorithm is applied to the filtered Pareto front core to partition it into clusters, compute the centroids, and then determine the medoids — real representative data points. Two experiments were conducted on three-dimensional Pareto front datasets (1226 and 2514 core points after filtering). As a result of applying the proposed approach, a partition into 10 clusters was achieved. It was found that after filtering, the proportion of noise points was less than 1 % of the total number of solutions. The filtering step significantly reduced the metric assessing the quality of cluster centers, with only a moderate increase in the total clustering time. A small discrepancy between centroids and their corresponding medoids indicates the high representativeness of the resulting clusters. The proposed hybrid method, combining Ordering Points to Identify the Clustering Structure and k-means algorithms, requires the adjustment of only two parameters and automatically adapts to nonlinear densities and input data scales. The scope of this method can be extended to any multi-objective optimization problems solved through the construction and analysis of the Pareto front, including engineering optimization, logistics, energy systems, and financial modeling. In the future, the approach may be enhanced by integrating adaptive mechanisms for automatic determination of optimal algorithm parameters, as well as dynamically changing multi-objective problem settings.
Keywords: clustering, Ordering Points to Identify the Clustering Structure, k-means, multi-objective optimization
References
References
1. Zack Yu.A. The set of Pareto efficiency criteria presented stochastic and fuzzy data. Artificial Intelligence and Decision Making, 2014, no. 2, pp. 89–101. (in Russian)
2. Nogin V.D. The Pareto Set and Principle. St. Petersburg, Izdatelsko-Poligraficheskaya Assotsiatsiya Vysshikh Uchebnykh Zavedeniy, 2022, 110 p. (in Russian)
3. Brester Ch.Yu., Stanovov V. V., Semenkina O. E., Semenkin E. S. About the use of evolutionary algorithms in big data analysis. Artificial Intelligence and Decision Making, 2017, no. 3, pp. 82–93. (in Russian)
4. Garey M.R., Johnson D.S. Computers and Intractability: A Guide to the Theory of NP-Completeness. W.H. Freeman and Company, 1979, 340 p.
5. Zitzler E., Deb K., Thiele L. Comparison of multiobjective evolutionary algorithms: empirical results. Evolutionary Computation, 2000, vol. 8, no. 2, pp. 173–195. https://doi.org/10.1162/106365600568202
6. Vatutin E.I.Solving Discrete Combinatorial Optimization Problems Using Heuristic Methods: Methodological Guidelines for Laboratory and Practical Works in the Course «Fundamentals of Combinatorial Optimization». Kursk, South-West State University Publ., 2016, 30 p. (in Russian)
7. Gladkov L.A., Kureichik V.V., Kureichik V.M. Genetic Algorithms. Moscow, Fizmatlit Publ., 2010, 368 p. (in Russian)
8. Wirsansky E. Hands-On Genetic Algorithms with Python: Applying Genetic Algorithms to Solve Real-World Deep Learning and Artificial Intelligence Problems. Packt Publishing, 2020, 346 p.
9. Everitt B.S., Landau S., Leese M., Stahl D. Cluster Analysis. John Wiley & Sons, 2011, 346 p.
10. Klinov D.A., Grigorian K.A. Development of a method for user segmentation using clustering algorithms and advanced analytics. Russian Digital Libraries Journal, 2022, vol. 25, no. 2, pp. 137–147. (in Russian). https://doi.org/10.26907/1562-5419-2022-25-2-137-147
11. Bondarenko I.B., Gatchin Y.A., Geranichev V.N. Synthesis of optimal artificial neural networks by modified genetic algorithm. Scientific and Technical Journal of Information Technologies, Mechanics and Optics, 2012, no. 2 (78), pp. 51–55. (in Russian)
12. Barsegyan A.A. Data and Process Analysis. St. Petersburg, BHV-Peterburg, 2009, 512 p. (in Russian)
13. Semerikov A.V., Glazyrin M.A. Clustering university students by hierarchical and KMeans methods. IT Arctica, 2021, no. 3, pp. 41–58. (in Russian)
14. Kislyakov A.N., Polyakov S.V. Hierarchical clustering methods in a task to find abnormal observations based on groups with broken symmetry. Administrative Consulting, 2020, no. 5 (137), pp 116–127. (in Russian). https://doi.org/10.22394/1726-1139-2020-5-116-127
15. Ivanov A.A. Data clustering based on a Markov chain using the DBSCAN algorithm // Novye Informatsionnye Tekhnologii v Avtomatizirovannykh Sistemakh, 2018, no. 21, pp. 315–319. (in Russian)
16. Trushkova K.N., Kalaida V.T. The clusterization of satellite images of cloudy fields on basis of algorithm DBSCAN. Izvestiya Vuzov. Fizika, 2013, vol. 56, no. 8-3, pp. 356–358. (in Russian)
17. Murphy K.P. Machine Learning: A Probabilistic Perspective. The MIT Press, 2012, 1104 p.
18. Garafiev I.Z.,Garafieva G.I. Clustering engineering vacancies: K-means and gaussian mixture model. Management Accounting, 2024, no. 9, pp. 234–240. (in Russian)
19. Chung F.R.K. Spectral Graph Theory. American Mathematical Society, 1997, 207 p.
20. Barinov A.E., Zakharov A.A., Zhiznyakov A.L. Spectral clustering algorithm with constraints for human face extraction in images. Dinamika Sistem, Mekhanizmov i Mashin, 2016, no. 4, pp. 222–228. (in Russian)
21. Pylov P.A., Protodyakonov A.V. Spectral clustering for image segmentation. Innovatsii. Nauka. Obrazovanie, 2020, no.23, pp. 274–277. (in Russian)
22. Han J., Kamber M., Pei J. Data Mining: Concepts and Techniques. Morgan Kaufmann, 2011, 744 p.
23. Ankerst M., Breunig M.M., Kriegel H.P., Sander J. OPTICS: ordering points to identify the clustering structure. Proc. of the 1999 ACM SIGMOD International Conference on Management of Data, 1999, pp. 49–60. https://doi.org/10.1145/304182.304187
24. Popova O.A. Analysis and selection of a method of clusterization of short length text documents for implementation in the university recommender system module. XXI Century: Resumes of the Past and Challenges of the Present Plus, 2023,vol. 12,no. 4 (64),pp. 89–102.(in Russian)
25. Yurtaev A.G., Stepanov M.F. Solution to the problem of selecting a component base in the design of electronic avionics blocks. Mathematical Methods in Technologies and Technics, 2025,no. 2,pp. 170–174.(in Russian)

