Papers
arxiv:2507.12142

RiemannLoRA: A Unified Riemannian Framework for Ambiguity-Free LoRA Optimization

Published on Jul 16
· Submitted by ai-alanov on Jul 18
Authors:
,
,
,
,
,

Abstract

RiemannLoRA addresses initialization and overparametrization in LoRA by treating LoRA matrices as a smooth manifold, improving convergence speed and performance in LLMs and diffusion models.

AI-generated summary

Low-Rank Adaptation (LoRA) has become a widely adopted standard for parameter-efficient fine-tuning of large language models (LLMs), significantly reducing memory and computational demands. However, challenges remain, including finding optimal initialization strategies or mitigating overparametrization in low-rank matrix factorization. In this work, we propose a novel approach that addresses both of the challenges simultaneously within a unified framework. Our method treats a set of fixed-rank LoRA matrices as a smooth manifold. Considering adapters as elements on this manifold removes overparametrization, while determining the direction of the fastest loss decrease along the manifold provides initialization. Special care is taken to obtain numerically stable and computationally efficient implementation of our method, using best practices from numerical linear algebra and Riemannian optimization. Experimental results on LLM and diffusion model architectures demonstrate that RiemannLoRA consistently improves both convergence speed and final performance over standard LoRA and its state-of-the-art modifications.

Community

Paper author Paper submitter

We propose RiemannLoRA that improves Low-Rank Adaptation for efficient fine-tuning of large models by treating LoRA matrices as a smooth manifold, addressing overparametrization and initialization challenges through Riemannian optimization, leading to faster convergence and better performance than standard LoRA methods.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2507.12142 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2507.12142 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2507.12142 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.