Empirical Inference Perceiving Systems Conference Paper 2024

Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization

Thumb ticker sm weiyang photo
Empirical Inference
  • Postdoctoral Researcher
Thumb ticker sm photo yao
Perceiving Systems
  • Guest Scientist
Thumb ticker sm 84b79ecc0b10a9579306eaa7c8037a
Perceiving Systems
  • Guest Scientist
no image
Embodied Vision
Thumb ticker sm img 4026
Perceiving Systems
  • Doctoral Researcher
Thumb ticker sm zhenliu
Perceiving Systems
no image
Autonomous Vision
Thumb ticker sm zyro image  8
Perceiving Systems
  • Guest Scientist
Thumb ticker sm headshot2021
Perceiving Systems
Director
Thumb ticker sm l1170153
Empirical Inference
  • Director
Thumb xxl comp lora teaser

Large foundation models are becoming ubiquitous, but training them from scratch is prohibitively expensive. Thus, efficiently adapting these powerful models to downstream tasks is increasingly important. In this paper, we study a principled finetuning paradigm -- Orthogonal Finetuning (OFT) -- for downstream task adaptation. Despite demonstrating good generalizability, OFT still uses a fairly large number of trainable parameters due to the high dimensionality of orthogonal matrices. To address this, we start by examining OFT from an information transmission perspective, and then identify a few key desiderata that enable better parameter-efficiency. Inspired by how the Cooley-Tukey fast Fourier transform algorithm enables efficient information transmission, we propose an efficient orthogonal parameterization using butterfly structures. We apply this parameterization to OFT, creating a novel parameter-efficient finetuning method, called Orthogonal Butterfly (BOFT). By subsuming OFT as a special case, BOFT introduces a generalized orthogonal finetuning framework. Finally, we conduct an extensive empirical study of adapting large vision transformers, large language models, and text-to-image diffusion models to various downstream tasks in vision and language.

Author(s): Weiyang Liu and Zeju Qiu and Yao Feng and Yuliang Xiu and Yuxuan Xue and Longhui Yu and Haiwen Feng and Zhen Liu and Juyeon Heo and Songyou Peng and Yandong Wen and Michael J. Black and Adrian Weller and Bernhard Schölkopf
Book Title: Proceedings of the Twelfth International Conference on Learning Representations (ICLR)
Year: 2024
Month: May
Bibtex Type: Conference Paper (inproceedings)
Event Name: The Twelfth International Conference on Learning Representations
Event Place: Vienna, Austria
State: Published
URL: https://openreview.net/forum?id=7NzgkEdGyr
Electronic Archiving: grant_archive
Links:

BibTex

@inproceedings{boft,
  title = {Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization},
  booktitle = {Proceedings of the Twelfth International Conference on Learning Representations (ICLR)},
  abstract = {Large foundation models are becoming ubiquitous, but training them from scratch is prohibitively expensive. Thus, efficiently adapting these powerful models to downstream tasks is increasingly important. In this paper, we study a principled finetuning paradigm -- Orthogonal Finetuning (OFT) -- for downstream task adaptation. Despite demonstrating good generalizability, OFT still uses a fairly large number of trainable parameters due to the high dimensionality of orthogonal matrices. To address this, we start by examining OFT from an information transmission perspective, and then identify a few key desiderata that enable better parameter-efficiency. Inspired by how the Cooley-Tukey fast Fourier transform algorithm enables efficient information transmission, we propose an efficient orthogonal parameterization using butterfly structures. We apply this parameterization to OFT, creating a novel parameter-efficient finetuning method, called Orthogonal Butterfly (BOFT). By subsuming OFT as a special case, BOFT introduces a generalized orthogonal finetuning framework. Finally, we conduct an extensive empirical study of adapting large vision transformers, large language models, and text-to-image diffusion models to various downstream tasks in vision and language.},
  month = may,
  year = {2024},
  slug = {boft},
  author = {Liu, Weiyang and Qiu, Zeju and Feng, Yao and Xiu, Yuliang and Xue, Yuxuan and Yu, Longhui and Feng, Haiwen and Liu, Zhen and Heo, Juyeon and Peng, Songyou and Wen, Yandong and Black, Michael J. and Weller, Adrian and Sch{\"o}lkopf, Bernhard},
  url = {https://openreview.net/forum?id=7NzgkEdGyr},
  month_numeric = {5}
}