[go: up one dir, main page]

Leena C Vankadara

Lecturer (Assistant Professor) @ Gatsby Unit, UCL.

l.vankadara [AT] ucl.ac.uk

Short Bio

I am a Lecturer (US Assistant Professor equivalent) at the Gatsby Unit, University College London. Previously, I was an Applied Scientist at the AGI Foundations Lab at Amazon, where I worked on fundamental research on the theory and science of scaling large language models. I earned my Ph.D. from the International Max Planck Research School for Intelligent Systems with Debarghya Ghoshdastidar at the Theoretical Foundations of Artificial Intelligence group, Technical University of Munich, and Ulrike von Luxburg at the Theory of Machine Learning group, University of Tuebingen. My thesis received the Wilhelm Schickard Dissertation Award (Outstanding Dissertation). During my Ph.D., I also spent 6 months at the Causality Lab of Amazon Research.

My research aims to develop efficient, reliable, and trustworthy machine learning models by understanding the theoretical foundations of deep learning and causal learning. Deep learning theory and causality present some of the most fascinating challenges in modern machine learning. Classical theories of statistical learning and optimization often struggle to explain the empirical success of deep learning algorithms or even provide reliable guidance on their behaviour. As models continue to scale to unprecedented levels, revealing emergent abilities, we need fundamentally new theories to explain their behavior. To tackle some of these questions, my research aims to i) develop a unified theory of scaling and ii) understand the learning dynamics and generalization in deep learning. Meanwhile, Causality remains in its nascency, with formal mechanisms still emerging. My research will address open problems in causality, including establishing robust formalisms for causal reasoning and developing frameworks to evaluate causal models.

Recent News
  • [27/09/2025] Our paper On the Surprising Effectiveness of Large Learning Rates under Standard Width Scaling has been accepted to NeurIPS 2025 (Spotlight)!. Paper
  • [Upcoming 03/2026] I will be giving a talk at the Modern and Emerging Phenomena in Machine Learning workshop, Oberwolfach, Germany.
  • [Upcoming 14/12/2025] I will be giving a tutorial at CAMSAP, Dominican Republic, together with Prof. Volkan Cevher (EPFL) on Training Neural Networks at Any Scale.
  • [Upcoming 05/11/2025] Excited to be giving a keynote talk at the UCL Neuro AI Annual Conference, London, UK.
  • [07/2025] Excited to be giving a keynote talk at the Women in Machine Learning (WiML) Symposium @ ICML 2025, Vancouver, Canada.
  • [07/2025] Excited to be giving a keynote talk at pre-ICML London 2025. Event
  • [07/2025] Excited to be giving a tutorial together with Prof. Volkan Cevher (EPFL) on Training Neural Networks at Any Scale at ICML 2025, Vancouver, Canada. Recording · Slides
  • [10/06/2025] Gave a talk at the Gatsby Tri‑centre Annual Meeting, London, UK.
  • [01/11/2024] I am thrilled to announce that I will be joining the Gatsby Computational Neuroscience Unit at UCL as a Lecturer (Assistant Professor) in Feb 2025. This new role marks an exciting chapter in my academic career. Stay tuned for updates on my upcoming courses, research projects, and other activities at the Gatsby Unit!

  • [25/07/2024] Two papers have been accepted to NeurIPS 2024! Very excited to be presenting our papers: "On Feature Learning in Structured State Space Models" and "mup2: Effective Sharpness Aware Minimization Requires Layerwise Perturbation Scaling" at the conference in December!
  • [24/09/2024] I am delighted to announce that I will be joining the Gatsby Computational Neuroscience Unit at UCL as a Lecturer (Assistant Professor) in February 2025!
  • [24/09/2024] Our paper A Consistent Estimator for Confounding Strength has been accepted by the Mathematical Statistics and Learning Journal!
  • [Upcoming 14/10/2024] Excited to be giving a talk on "Towards a theory of scaling" at the Theory and Practice of Deep Learning, IPAM, UCLA, Los Angeles, U.S.A!
  • [Upcoming 14/12/2024] Will be giving a talk on "Towards a theory of scaling" at the International Joint Conference CFE-CMStatistics, London, UK.
  • [25/07/2024] Excited to announce that our work "On Feature Learning in Structured State Space Models" has been accepted to the ICML NGSM Workshop 2024 as a Spotlight presentation, and it was also the Runner-up for Best Paper Award!
  • [25/07/2024] We will be presenting our paper "Effective Sharpness Aware Minimization Requires Layerwise Perturbation Scaling" at the ICML HiLD Workshop 2024.
  • [10/07/2024] Giving a tutorial on "Scaling and Reliability Foundations in Machine Learning" at the IEEE International Symposium on Information Theory, Athens, Greece!
  • [20/06/2024] Giving a talk on "Scaling theory for sharpness-aware minimization" at INRIA-LILLE, Nord Europe, Villeneuve-d'Ascq, France.
  • [06/05/2024] Our work "Explaining Kernel Clustering via Decision Trees" has been accepted to ICLR 2024!
  • [02/05/2024] We will be presenting "Self-Compatibility: Evaluating Causal Discovery without Ground Truth" at AISTATS 2024!
  • [05/02/2024] Giving a talk on "Interpolation and Regularization for Causal Learning" at the Gatsby Computational Neuroscience Unit, London, UK.
  • [01/02/2024] Moved to the AGI Foundations team at Amazon as an Applied Scientist II. Will be working on theory and science of scaling large language models.
  • [12/12/2023] Giving a talk on "Is Memorization Compatible with Causal Learning. The Case of High-Dimensional Linear Regression" at the International Joint Conference CFE-CMStatistics, Berlin, Germany.
  • [01/01/2023] Started a new position in the Causality lab at Amazon Research as an Applied Scientist II.
  • [21/12/2022] Submitted my PhD Thesis!
  • [02/12/2022] Luca Rendsburg and I will be presenting our work on Interpolation and Regularization for Causal Learning. at NeurIPS 2022 in New Orleans, USA.
  • [03/11/2022] New Preprint: A Consistent Estimator for Confounding Strength.
  • [15/09/2022] Paper: Interpolation and Regularization for Causal Learning. accepted to NeurIPS 2022.
  • [14/09/2022] Giving a talk on Causal properties of interpolating estimators in the Department of Mathematical Statistics, Technical University of Munich.
  • [27/05/2022] I will be in London for two weeks attending the annual meeting of the Institute of Mathematical Statistics and COLT 2022!
  • [19/05/2022] Giving a talk on "Interpolation and Regularization for Causal Learning" at the Women in Data Science conference in Chemnitz.
  • [27/04/2022] Mahalakshmi Sabanayagam will be presenting our work on Graphon based Clustering and Testing of Networks: Algorithms and Theory at ICLR 2022!
  • [24/03/2022] Giving a talk on "Kernels for High-Dimensional Large-Scale Clustering" at the SIAM Conference on Imaging Science.
Advisees/Interns
    Current
    • Luke Hayward, University of Oxford
    • Anish Dhir, Imperial College, London
    • Leello Tadesse Dadi , École Polytechnique Fédérale de Lausanne (EPFL)
    • Zhenyu Zhu, École Polytechnique Fédérale de Lausanne (EPFL)
    • Moritz Haas, University of Tuebingen
    • Vaclav Voracek, University of Tuebingen
    • Valentyn Boreiko , University of Tuebingen
    Previous
    • Jin Xu, University of Oxford
    • Alexander Meulemans, ETH, Zurich
    • Maximilian Fleissner, Technical University of Munich
    • Mahalakshmi Sabanayagam, Technical University of Munich
    • Margareta Schlüter, University of Tuebingen
Publications

Most recent publications on Google Scholar.
indicates equal contribution.

  • Selected
  • All

On the surprising effectivess of large learning rates under standard width scaling

Moritz Haas, Sebastian Bordt, Ulrike von Luxburg, Leena C Vankadara*

NeurIPS 2025 (Spotlight Presentation)

On Feature Learning in Structured State Space Sequence Models

Leena C Vankadara*, Jin Xu*, Moritz Haas, Volkan Cevher

NeurIPS 2024 & ICML NGSM Workshop 2024 (Spotlight; Runner up for the best paper award)

Effective Sharpness Aware Minimization Requires Layerwise Perturbation Scaling

Moritz Haas, Jin Xu, Volkan Cevher, Leena C Vankadara

NeurIPS 2024 & ICML HiLD Workshop 2024

Self-compatibility: Evaluating causal discovery without ground truth

Philipp M Faller, Leena C Vankadara, Atalanti A Mastakouri, Francesco Locatello, Dominik Janzing

AISTATS 2024

Reinterpreting causal discovery as the task of predicting unobserved joint statistics

Dominik Janzing, Philipp M. Faller, Leena C Vankadara

A preprint: Arxiv'2023

Explaining Kernel Clustering via Decision Trees

Maximilian Fleissner, Leena C Vankadara, Debarghya Ghoshdastidar

ICLR 2024

A Consistent Estimator for Confounding Strength

Luca Rendsburg, Leena C Vankadara, Debarghya Ghosdastidar, Ulrike von Luxburg

A preprint: Arxiv'2022

Interpolation and Regularization for Causal Learning

Leena C Vankadara, Luca Rendsburg, Ulrike von Luxburg, Debarghya Ghoshdastidar

NeurIPS'22: Neural Information Processing Systems. 2022.

Causal Forecasting - Generalization Bounds for Autoregressive Models

Leena C Vankadara, Philipp Michael Faller, Mila Hardt, Lenon Minorics, Debarghya Ghoshdastidar, Dominik Janzing

UAI'22: Uncertainity in Artificial Intelligence. 2022.

Recovery Guarantees for Kernel-based Clustering under Non-parametric Mixture Models

Leena C Vankadara, Sebastian Bordt, Ulrike von Luxburg, Debarghya Ghoshdastidar

AISTATS'21: Artificial Intelligence and Statistics. 2021 (Oral presentation; 3% of all submissions)

On the optimality of kernels for high-dimensional clustering

Leena C Vankadara, Debarghya Ghoshdastidar

AISTATS'20: Artificial Intelligence and Statistics. 2020

On the surprising effectivess of large learning rates under standard width scaling

Moritz Haas, Sebastian Bordt, Ulrike von Luxburg, Leena C Vankadara*

NeurIPS 2025 (Spotlight Presentation)

On Feature Learning in Structured State Space Sequence Models

Leena C Vankadara*, Jin Xu*, Moritz Haas, Volkan Cevher

NeurIPS 2024 & ICML NGSM Workshop 2024 (Spotlight; Runner up for the best paper award)

Effective Sharpness Aware Minimization Requires Layerwise Perturbation Scaling

Moritz Haas, Jin Xu, Volkan Cevher, Leena C Vankadara

NeurIPS 2024 & ICML HiLD Workshop 2024

Self-compatibility: Evaluating causal discovery without ground truth

Philipp M Faller, Leena C Vankadara, Atalanti A Mastakouri, Francesco Locatello, Dominik Janzing

AISTATS 2024

Reinterpreting causal discovery as the task of predicting unobserved joint statistics

Dominik Janzing, Philipp M. Faller, Leena C Vankadara

A preprint: Arxiv'2023

Explaining Kernel Clustering via Decision Trees

Maximilian Fleissner, Leena C Vankadara, Debarghya Ghoshdastidar

ICLR 2024

A Consistent Estimator for Confounding Strength

Luca Rendsburg, Leena C Vankadara, Debarghya Ghosdastidar, Ulrike von Luxburg

A preprint: Arxiv'2022

Interpolation and Regularization for Causal Learning

Leena C Vankadara, Luca Rendsburg, Ulrike von Luxburg, Debarghya Ghoshdastidar

NeurIPS'22: Neural Information Processing Systems. 2022.

Causal Forecasting - Generalization Bounds for Autoregressive Models

Leena C Vankadara, Philipp Michael Faller, Mila Hardt, Lenon Minorics, Debarghya Ghoshdastidar, Dominik Janzing

UAI'22: Uncertainity in Artificial Intelligence. 2022.

Graphon based Clustering and Testing of Networks - Algorithms and Theory

Mahalakshmi Sabanayagam, Leena C Vankadara, Debarghya Ghoshdastidar

ICLR'22: International Conference on Learning Representations. 2022

Learning Theory Can (Sometimes) Explain Generalisation in Graph Neural Networks

Pascal Esser, Leena C Vankadara, Debarghya Ghoshdastidar

NeurIPS'21: Neural Information Processing Systems. 2021

Recovery Guarantees for Kernel-based Clustering under Non-parametric Mixture Models

Leena C Vankadara, Sebastian Bordt, Ulrike von Luxburg, Debarghya Ghoshdastidar

AISTATS'21: Artificial Intelligence and Statistics. 2021 (Oral presentation; 3% of all submissions)

Insights into Ordinal Embedding Algorithms - A Systematic Evaluation

Leena C Vankadara, Michael Lohaus, Siavash Haghiri, Faiz Ul Wahab, Ulrike von Luxburg

JMLR. 2023

On the optimality of kernels for high-dimensional clustering

Leena C Vankadara, Debarghya Ghoshdastidar

AISTATS'20: Artificial Intelligence and Statistics. 2020

Measures of distortion for machine learning

Leena C Vankadara, Ulrike von Luxburg

NeurIPS'18: Neural Information Processing Systems. 2018

Vitæ
  • Gatsby Unit, UCL Starting Feb 2025
    Lecturer (Assistant Professor)
  • Amazon Research January 2024 - now
    Applied Scientist II
    AGI Foundations, Amazon
  • Amazon Research January 2023 - now
    Applied Scientist II
    Causality Lab
  • IMPRS-IS 2018 - 2022
    Ph.D. Candidate
    Theory of Machine Learning Group
  • Amazon Research Fall 2020
    Applied Scientist, Intern
    Causality Lab
  • Max Planck Institute (MPI-IS) June - December 2018
    Research Scientist, Intern
    Statistical Learning Theory Group
  • University of Tuebingen April 2017 – June 2018
    Teaching Assistant
    Theory of Machine learning
  • University of Tuebingen 2017
    Master's Thesis
    Theory of Machine learning group
  • University of Hamburg 2014 - 2017
    Master's Student
    Masters in Machine learning