Bermak, Amine
Systematic Weight Evaluation for Pruning Large Language Models: Enhancing Performance and Sustainability
Islam, Ashhadul, Belhaouari, Samir Brahim, Bermak, Amine
Impact of Compression on Model Performance: Through comprehensive experiments, the study demonstrates that moderate pruning can enhance model efficiency, but excessive compression leads to substantial performance degradation in both language and multimodal models. Sustainable AI Development: The findings emphasize the need for optimized AI models to reduce the environmental impact, addressing critical issues like carbon footprint, electricity, and water consumption associated with training and deploying large-scale AI systems. Systematic Weight Evaluation for Pruning Large Language Models: Enhancing Performance and Sustainability Ashhadul Islam a, Samir Brahim Belhaouari a, Amine Bermak a a College Of Science & Engineering, Hamad Bin Khalifa University, Education City, Doha, 34110, QatarAbstract The exponential growth of large language models (LLMs) like ChatGPT has revolutionized artificial intelligence, offering unprecedented capabilities in natural language processing. However, the extensive computational resources required for training these models have significant environmental implications, including high carbon emissions, energy consumption, and water usage. This research presents a novel approach to LLM pruning, focusing on the systematic evaluation of individual weight importance throughout the training process. By monitoring parameter evolution over time, we propose a method that effectively reduces model size without compromising performance. Extensive experiments with both a scaled-down LLM and a large multimodal model reveal that moderate pruning enhances efficiency and reduces loss, while excessive pruning drastically deteriorates model performance. These findings highlight the critical need for optimized AI models to ensure sustainable development, balancing technological advancement with environmental responsibility.
Deep Exemplar-based Video Colorization
Zhang, Bo, He, Mingming, Liao, Jing, Sander, Pedro V., Yuan, Lu, Bermak, Amine, Chen, Dong
This paper presents the first end-to-end network for exemplar-based video colorization. The main challenge is to achieve temporal consistency while remaining faithful to the reference style. To address this issue, we introduce a recurrent framework that unifies the semantic correspondence and color propagation steps. Both steps allow a provided reference image to guide the colorization of every frame, thus reducing accumulated propagation errors. Video frames are colorized in sequence based on the colorization history, and its coherency is further enforced by the temporal consistency loss. All of these components, learned end-to-end, help produce realistic videos with good temporal stability. Experiments show our result is superior to the state-of-the-art methods both quantitatively and qualitatively.