Doctoral thesis (Dissertations and theses)
Performance Prediction Models for Deep Learning: A Graph Neural Network and Large Language Model Approach
PANNER SELVAM, Karthick
2025
 

Files


Full Text
Karthick_PhD_Thesis.pdf
Author postprint (4.87 MB) Creative Commons License - Attribution, Non-Commercial
Download

All documents in ORBilu are protected by a user license.

Send to



Details



Keywords :
Graph Neural Networks; Large Language Models; Machine Learning for Systems; Performance Prediction Models
Abstract :
[en] In this thesis, we developed an advanced performance prediction model to estimate critical metrics of Deep Learning (DL) models, such as latency, memory consumption, and energy usage. These models are designed to support neural architecture search and efficient cloud deployment. DL has transformed domains such as computer vision, natural language processing, climate modeling, and scientific computing. However, the increasing complexity of DL models introduces significant computational demands that require efficient hardware utilization and resource allocation. Accurate prediction of performance metrics is essential for optimizing hardware-specific compilers, enabling cost-effective cloud deployments, and minimizing environmental impacts. To address these challenges, we present a comprehensive framework for performance prediction in this thesis. The work begins with a systematic benchmarking study of DL models, highlighting computational bottlenecks and establishing the necessity of performance prediction as a foundation for further development. We introduce a Graph Neural Network (GNN) based performance prediction model capable of analyzing DL models from various software frameworks, including PyTorch and TensorFlow. This model predicts performance metrics and recommends NVIDIA multi-GPU instance profiles for efficient deployment. Building on this, we propose a semi-supervised performance prediction approach that leverages unlabeled data to accelerate training convergence. Using a graph autoencoder for unsupervised learning, we generate high-quality embeddings that enhance supervised training, leading to faster and more accurate predictions. For Large Language Models (LLMs), which present unique challenges due to their extensive nodes and edges, we proposed a tree-based performance prediction model. This method significantly improves inference speed compared to traditional GNN-based techniques, making it particularly suitable for complex LLM architectures. Finally, we explore multimodal learning by combining LLM with GNN to create a hybrid performance prediction model. This model quickly adapts to new hardware environments with sparse training samples, leveraging a novel three-stage training strategy to effectively integrate GNN and LLM for quick adaptation.
Research center :
Interdisciplinary Centre for Security, Reliability and Trust (SnT) > SEDAN - Service and Data Management in Distributed Systems
Disciplines :
Computer science
Author, co-author :
PANNER SELVAM, Karthick ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SEDAN
Language :
English
Title :
Performance Prediction Models for Deep Learning: A Graph Neural Network and Large Language Model Approach
Defense date :
2025
Institution :
Unilu - University of Luxembourg, Luxembourg
Degree :
Docteur en Informatique (DIP_DOC_0006_B)
Jury member :
BRORSSON, Mats Håkan  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SEDAN
BOUVRY, Pascal ;  University of Luxembourg > Faculty of Science, Technology and Medicine (FSTM) > Department of Computer Science (DCS)
STATE, Radu  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SEDAN
VLASSOV, Vladimir;  KTH - Royal Institute of Technology
PHOTHILIMTHANA, Phitchaya Mangpo;  OpenAI
Focus Area :
Security, Reliability and Trust
European Projects :
H2020 - 955513 - MAELSTROM - MAchinE Learning for Scalable meTeoROlogy and cliMate
FnR Project :
FNR15092355 - MAELSTROM - Machine Learning For Scalable Meteorology And Climate, 2020 (01/04/2021-31/03/2024) - Mats Brorsson
Name of the research project :
U-AGR-8013 - INTER/EuroHPC/20/15077233/MAELSTROM - BRORSSON Mats Hakan
Funders :
EuroHPC JU
FNR - Luxembourg National Research Fund
European Union
Funding number :
955513; 15092355
Funding text :
This work has been done in the context of the MAELSTROM project, which has received funding from the European High-Performance Computing Joint Undertaking (JU) under grant agreement No 955513. The JU receives support from the European Union’s Horizon 2020 research and innovation program and United Kingdom, Germany, Italy, Switzerland, Norway, and in Luxembourg by the Luxembourg National Research Fund (FNR) under contract number 15092355.
Available on ORBilu :
since 27 March 2025

Statistics


Number of views
258 (16 by Unilu)
Number of downloads
451 (12 by Unilu)

Bibliography


Similar publications



Contact ORBilu