While overparameterization in machine learning models offers great benefits in terms of optimization and generalization, it also leads to increased computational requirements as model sizes grow. In this work, we show that leveraging inherent low-dimensional structure within the model parameter updates, we can reap the benefits of overparameterization without the computational burden. In practice, we demonstrate the effectiveness of this approach for deep low-rank matrix completion as well as fine-tuning language models. For theory of deep overparameterized low-rank matrix recovery, we show that the learning dynamics of each weight matrix are confined to an invariant low-dimensional subspace. Consequently, we can construct and train compact, highly compressed factorizations possessing the same benefits as their overparameterized counterparts. For language model fine-tuning, we introduce a method called “Deep LoRA”, which improves the existing low-rank adaptation (LoRA) technique, leading to reduced overfitting and a simplified hyperparameter setup, all while maintaining comparable efficiency. The effectiveness of Deep LoRA is validated through its performance on natural language understanding tasks, particularly when fine-tuning with a limited number of samples.
Human-Centered Explainable AI for Healthcare
June 4, 2024
3:24 pm
Catarina Barata
Catarina Barata holds a Msc Degree in Biomedical Engineering and a PhD in Electrical and Computer Engineering (Instituto Superior Técnico - IST - 2011 and 2017 respectively). In the Fall of 2022, she was a Visiting Scholar at Carnegie Mellon University. Presently, she is a tenure-track Assistant Professor at the ECE Department of IST and a Researcher at Institute for Systems and Robotics (ISR), where she is a member of the Computer and Robot Vision Laboratory (VisLab). Her main research interests are in the interface between machine learning, computer vision, and healthcare, where she has been collaborating and leading various projects together with hospitals and other healthcare institutions. An example is her work on the discovery of therapeutic biomarkers for melanoma, for which she received a Google Research Award in 2021.Seminários
Últimos seminários
Unlocking Latent Discourse Translation in LLMs Through Quality-Aware Decoding
June 17, 2025Large language models (LLMs) have emerged as strong contenders in machine translation. Yet, they often fall behind specialized neural machine…
Speech as a Biomarker for Disease Detection
May 20, 2025Today’s overburdened health systems face numerous challenges, exacerbated by an aging population. Speech emerges as a ubiquitous biomarker with strong…
Enhancing Uncertainty Estimation in Neural Networks
May 6, 2025Neural networks are often overconfident about their predictions, which undermines their reliability and trustworthiness. In this presentation, I will present…
Improving Evaluation Metrics for Vision-and-Language Models
April 22, 2025Evaluating image captions is essential for ensuring both linguistic fluency and accurate semantic alignment with visual content. While reference-free metrics…



