Thèse de doctorat (Mémoires et thèses)
Maintenance optimization in Industry 4.0: A Deep Reinforcement Learning Approach to Sustainable Policy Development
RUIZ RODRIGUEZ, Marcelo Luis
2024
 

Documents


Texte intégral
ThesisFinal.pdf
Postprint Auteur (9.31 MB)
Télécharger

Tous les documents dans ORBilu sont protégés par une licence d'utilisation.

Envoyer vers



Détails



Mots-clés :
Reinforcement Learning; Maintenance; Industry 4.0; Manufacturing; Sustainability
Résumé :
[en] Effective maintenance planning and scheduling are essential for manufacturing companies to prevent machine breakdowns and maximize uptime and production. Furthermore, these policies must be in alignment with the principles of environmental integrity and social responsibility. The development of sustainable policies presents several challenges. These include the need to balance economic, environmental, and social aspects, as well as to address uncertainties such as unexpected failures and variable time-to-repair (TTR). This thesis, conducted in partnership with Cebi, an electromechanical component design and manufacturing company, addresses the challenge of developing sustainable maintenance policies in the face of uncertainty. To this end, we propose a Deep Reinforcement Learning (DRL)-based approach for predictive maintenance, which we compare with traditional maintenance policies such as corrective, preventive, and condition-based maintenance, and evaluate against diverse methods based on metaheuristics and rule-based approaches. As a first contribution, we conducted a study of the impact of categorized levels of uncertainty in a manufacturing environment on the failure distribution and time to repair for maintenance policies. We evaluated the performance of DRL, genetic algorithm-based simheuristic (GA-S), and rule-based (DR) decision-making systems in terms of mean time to repair (MTTR), machine uptime, and computational efficiency. The work was conducted in simulated scenarios with different levels of uncertainty and also considering a real manufacturing use case. The experimental results show that DRL shows exceptional adaptability to reduce MTTR, especially in the face of high uncertainty. GA-S outperforms DRL and DR in terms of total machine uptime, but not in terms of MTTR, when configured with high re-optimization frequencies (i.e., hourly re-optimization), but rapidly underperforms as the re-optimization frequency decreases. Furthermore, our study shows that GA-S is computationally expensive compared to DRL and DR policies. As a second contribution, we proposed to tackle the problem of maintenance scheduling on multi-component identical parallel machines performed by technicians. In this work, we proposed a multi-agent DRL approach to learn the maintenance policy under the uncertainty of multiple machine failures. This approach comprises DRL agents that partially observe the state of each machine to coordinate the decision-making in maintenance scheduling, resulting in the dynamic assignment of maintenance tasks to technicians (with different skills) over a set of machines. Experimental evaluation shows that our DRL-based maintenance policy outperforms classical maintenance policies such as Corrective Maintenance (CM) and Preventive Maintenance (PM) in terms of failure prevention and downtime, improving overall performance by $\approx 75\%$. Our last contribution proposed to optimize maintenance scheduling from an economic perspective (considering maintenance, breakdown, and downtime costs), an environmental perspective (considering the carbon footprint produced during production) and a social perspective (considering the fatigue experienced by technicians during maintenance activities). We propose an evolutionary multi-objective multi-agent deep Q-Learning (EvoDQN)-based approach, where multiple agents explore the preference space to maximize the hypervolume of these sustainable objectives. The results demonstrate the trade-offs between these objectives when compared to traditional maintenance policies such as condition-based maintenance (CBM) and CM, as well as different deep Q-network (DQN) policies trained with various preferences and a higher number of agents for our EvoDQN approach. Finally, our approach evaluated the production cycle, in which our method demonstrated superior performance, resulting in increased profitability within the system.
Centre de recherche :
Interdisciplinary Centre for Security, Reliability and Trust (SnT) > SerVal - Security, Reasoning & Validation
Disciplines :
Sciences informatiques
Auteur, co-auteur :
RUIZ RODRIGUEZ, Marcelo Luis ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SerVal
Langue du document :
Anglais
Titre :
Maintenance optimization in Industry 4.0: A Deep Reinforcement Learning Approach to Sustainable Policy Development
Date de soutenance :
21 novembre 2024
Institution :
Unilu - University of Luxembourg [The Faculty of Science, Technology and Medicine], Esch-sur-Alzette, Luxembourg
Intitulé du diplôme :
Docteur en Informatique (DIP_DOC_0006_B)
Promoteur :
LE TRAON, Yves ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT)
Président du jury :
NAVET, Nicolas ;  University of Luxembourg > Faculty of Science, Technology and Medicine (FSTM) > Department of Computer Science (DCS)
Membre du jury :
KUBLER, Sylvain ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SerVal
Bril El Haouzi, Hind;  UL - Université de Lorraine
Trentesaux, Damien;  UPHF - Université polytechnique des Hauts-de-France
Focus Area :
Computational Sciences
Organisme subsidiant :
FNR - Luxembourg National Research Fund
N° du Fonds :
16756339
Disponible sur ORBilu :
depuis le 28 novembre 2024

Statistiques


Nombre de vues
229 (dont 13 Unilu)
Nombre de téléchargements
188 (dont 8 Unilu)

Bibliographie


Publications similaires



Contacter ORBilu