publications
* denotes co-first authorship. Check also my Google Scholar profile.
2025
- arXivHow compositional generalization and creativity improve as diffusion models are trainedarXiv preprint, 2025
- ICLRUnraveling the latent hierarchical structure of language and images via diffusion modelsInternational Conference on Learning Representations, 2025
- ICLRLiNeS: Post-training layer scaling prevents forgetting and enhances model mergingInternational Coference on Learning Representations, 2025
- PNASA phase transition in diffusion models reveals the hierarchical nature of dataProceedings of the National Academy of Sciences, 2025
2024
- CVPRMulti-modal hallucination control by visual information groundingIEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024
- ICMLTask addition and weight disentanglement in closed-vocabulary modelsICML 2024 Efficient Systems for Foundation Models Workshop, 2024
- PhysRevXHow deep neural networks learn compositional data: The Random Hierarchy ModelPhysical Review X, 2024
- JSTATComputational complexity of deep learning: fundamental limitations and empirical phenomenaJournal of Statistical Mechanics: Theory and Experiment, 2024
2023
- ICMLWhat can be learnt with wide convolutional neural networks?International Conference on Machine Learning, PMLR, 2023
2021
- NeurIPSLocality defeats the curse of dimensionality in convolutional teacher-student scenariosAdvances in Neural Information Processing Systems, 2021
- NeurIPSRelative stability toward diffeomorphisms indicates performance in deep netsAdvances in Neural Information Processing Systems, 2021
2020
- ThesisSpectral analysis of infinitely wide convolutional neural networksMaster’s Thesis, Sorbonne Université and Politecnico di Torino, 2020