publications

* denotes equal contribution

preprints

  1. PLoP: Precise LoRA Placement for Efficient Finetuning of Large Models
    Soufiane Hayou, Nikhil Ghosh, and Bin Yu
    arXiv preprint, 2025

conference & journal articles

2025

  1. JMLR
    The Effect of SGD Batch Size on Autoencoder Learning: Sparsity, Sharpness, and Feature Learning
    Nikhil Ghosh, Spencer Frei, Wooseok Ha, and Bin Yu
    Journal of Machine Learning Research, 2025

2024

  1. NeuRIPS
    The Impact of Initialization on LoRA Finetuning Dynamics
    Soufiane Hayou, Nikhil Ghosh, and Bin Yu
    2024
  2. ICLR
    More is Better in Modern Machine Learning: when Infinite Overparameterization is Optimal and Overfitting is Obligatory
    James B. Simon, Dhruva Karkada, Nikhil Ghosh, and Mikhail Belkin
    2024

2023

  1. NeuRIPS
    Alternating Updates for Efficient Transformers
    Cenk Baykal, Dylan Cutler, Nishanth Dikkala, Nikhil Ghosh, Rina Panigrahy, and Xin Wang
    In Advances in Neural Information Processing Systems (NeurIPS), 2023
  2. SIMODS
    A Universal Trade-off Between the Model Size, Test Loss, and Training Loss of Linear Predictors
    Nikhil Ghosh and Mikhail Belkin
    SIAM Journal on Mathematics of Data Science (SIMODS), 2023
  3. ICLR
    Deconstructing Distributions: A Pointwise Framework of Learning
    Gal Kaplun*, Nikhil Ghosh*, Saurabh Garg, Boaz Barak, and Preetum Nakkiran
    In International Conference on Learning Representations (ICLR), 2023

2022

  1. ICLR
    The Three Stages of Learning Dynamics in High-Dimensional Kernel Methods
    Nikhil Ghosh, Song Mei, and Bin Yu
    In International Conference on Learning Representations (ICLR), 2022

2019

  1. NeuRIPS
    Landmark Ordinal Embedding
    Nikhil Ghosh, Yuxin Chen, and Yisong Yue
    In Advances in Neural Information Processing Systems (NeurIPS), 2019