Cora: Correspondence-aware image editing using few step diffusion
Abstract
Cora framework enhances image editing through correspondence-aware noise correction and interpolated attention maps, excelling in structure and texture preservation and generation.
Image editing is an important task in computer graphics, vision, and VFX, with recent diffusion-based methods achieving fast and high-quality results. However, edits requiring significant structural changes, such as non-rigid deformations, object modifications, or content generation, remain challenging. Existing few step editing approaches produce artifacts such as irrelevant texture or struggle to preserve key attributes of the source image (e.g., pose). We introduce Cora, a novel editing framework that addresses these limitations by introducing correspondence-aware noise correction and interpolated attention maps. Our method aligns textures and structures between the source and target images through semantic correspondence, enabling accurate texture transfer while generating new content when necessary. Cora offers control over the balance between content generation and preservation. Extensive experiments demonstrate that, quantitatively and qualitatively, Cora excels in maintaining structure, textures, and identity across diverse edits, including pose changes, object addition, and texture refinements. User studies confirm that Cora delivers superior results, outperforming alternatives.
Community
Cora is a fast image editing method that uses semantic correspondences to enable accurate edits in just 4 diffusion steps.
It also offers flexibility through attention and structure alignment, letting users control how much of the original image to preserve or change.
Website: cora-edit.github.io
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Structure-Preserving Zero-Shot Image Editing via Stage-Wise Latent Injection in Diffusion Models (2025)
- MDE-Edit: Masked Dual-Editing for Multi-Object Image Editing via Diffusion Models (2025)
- Tuning-Free Image Editing with Fidelity and Editability via Unified Latent Diffusion Model (2025)
- Improving Editability in Image Generation with Layer-wise Memory (2025)
- Beyond Editing Pairs: Fine-Grained Instructional Image Editing via Multi-Scale Learnable Regions (2025)
- Anchor Token Matching: Implicit Structure Locking for Training-free AR Image Editing (2025)
- Multi-turn Consistent Image Editing (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper