TR2024-157

Slaying the HyDRA: Parameter-Efficient Hyper Networks with Low-Displacement Rank Adaptation


    •  Chen, X., Wang, Y., Brand, M., Wang, P., Liu, J., Koike-Akino, T., "Slaying the HyDRA: Parameter-Efficient Hyper Networks with Low-Displacement Rank Adaptation", Advances in Neural Information Processing Systems (NeurIPS), December 2024.
      BibTeX TR2024-157 PDF
      • @inproceedings{Chen2024dec,
      • author = {Chen, Xiangyu and Wang, Ye and Brand, Matthew and Wang, Pu and Liu, Jing and Koike-Akino, Toshiaki}},
      • title = {Slaying the HyDRA: Parameter-Efficient Hyper Networks with Low-Displacement Rank Adaptation},
      • booktitle = {Advances in Neural Information Processing Systems (NeurIPS)},
      • year = 2024,
      • month = dec,
      • url = {https://www.merl.com/publications/TR2024-157}
      • }
  • MERL Contacts:
  • Research Areas:

    Artificial Intelligence, Machine Learning

Abstract:

Low-rank adaptation (LoRA) is a widely used parameter-efficient fine-tuning (PEFT) method to adapt large foundation models on downstream tasks. However, the weight updates are constrained to be low-rank structures, limiting their expressiveness. Alternatively, low-displacement rank (LDR)-based structured matrices are rank unrestricted, while requiring few parameters and supporting fast matrix- vector multiplication. We propose a new PEFT strategy to construct the weight updates with block-wise LDR matrices by sampling parameters from a hyper net- work framework. Our method, hyper low-displacement rank adaptation (HyDRA), offers high flexibility for choosing the size of a pool of trainable parameters, while not being restricted by the displacement rank. Our experiments demonstrate that the HyDRA can boost the classification accuracy by up to 3.4% and achieve two- fold improvement in parameter efficiency on an image classification benchmark compared with other PEFT variants.

 

  • Related Research Highlights