Balancing Performance, Efficiency and Robustness in Open-World Machine Learning via Evolutionary Multi-objective Model Compression

Javier Del Ser*, Aitor Martinez-Seras, Miren Nekane Bilbao, Jesus L. Lobo, Ibai Laña, Francisco Herrera

*Autor correspondiente de este trabajo

Producción científica: Capítulo del libro/informe/acta de congresoContribución a la conferenciarevisión exhaustiva

Resumen

When deploying machine learning models on resource-constrained hardware, reducing the memory footprint required by the model without compromising its performance is critical. Moreover, in open-world scenarios models often operate in dynamic and unpredictable environments where the data distribution evolves over time. Robust models can generalize well to unforeseen circumstances, including out-of-distribution inputs that may not have been encountered during the training phase. This adaptability is essential to handle the inherent variability of real-world data. This work formulates a multi-objective optimization problem that aims at optimizing the quantization resolution of the parameters of an already trained machine learning model based on three conflicting goals: maximizing the performance of the model on its designated learning task, minimizing the memory footprint of the compressed model, and enhancing its robustness against out-of-distribution data. Given the complexity of the resulting combinatorial optimization problem, we employ multi-objective evolutionary algorithms to efficiently obtain an approximation of the Pareto front balancing among the aforementioned objectives. Experiments with a randomized neural network compressed under the proposed formulation are run over several benchmark classification datasets. Different multi-objective solvers are employed to compare their effectiveness in terms of the convergence and diversity of their produced Pareto estimations. Additionally, we assess the achieved equilibrium between the three objectives against a floating-point implementation of the same model. Our experiments reveal that both the computational resources and the robustness of the model can be optimized via evolutionary quantization without significantly sacrificing its performance for the task at hand.

Idioma originalInglés
Título de la publicación alojada2024 International Joint Conference on Neural Networks, IJCNN 2024 - Proceedings
EditorialInstitute of Electrical and Electronics Engineers Inc.
ISBN (versión digital)9798350359312
DOI
EstadoPublicada - 2024
Evento2024 International Joint Conference on Neural Networks, IJCNN 2024 - Yokohama, Japón
Duración: 30 jun 20245 jul 2024

Serie de la publicación

NombreProceedings of the International Joint Conference on Neural Networks

Conferencia

Conferencia2024 International Joint Conference on Neural Networks, IJCNN 2024
País/TerritorioJapón
CiudadYokohama
Período30/06/245/07/24

Huella

Profundice en los temas de investigación de 'Balancing Performance, Efficiency and Robustness in Open-World Machine Learning via Evolutionary Multi-objective Model Compression'. En conjunto forman una huella única.

Citar esto