Abstract
Restoring fine-tuning to a breadth-first pipeline with mini-batch optimization and localized tuning parameters improves its effectiveness for model editing, outperforming state-of-the-art methods.
Fine-tuning, a foundational method for adapting large language models, has long been considered ineffective for model editing. Here, we challenge this belief, arguing that the reported failure arises not from the inherent limitation of fine-tuning itself, but from adapting it to the sequential nature of the editing task, a single-pass depth-first pipeline that optimizes each sample to convergence before moving on. While intuitive, this depth-first pipeline coupled with sample-wise updating over-optimizes each edit and induces interference across edits. Our controlled experiments reveal that simply restoring fine-tuning to the standard breadth-first (i.e., epoch-based) pipeline with mini-batch optimization substantially improves its effectiveness for model editing. Moreover, fine-tuning in editing also suffers from suboptimal tuning parameter locations inherited from prior methods. Through systematic analysis of tuning locations, we derive LocFT-BF, a simple and effective localized editing method built on the restored fine-tuning framework. Extensive experiments across diverse LLMs and datasets demonstrate that LocFT-BF outperforms state-of-the-art methods by large margins. Notably, to our knowledge, it is the first to sustain 100K edits and 72B-parameter models,10 x beyond prior practice, without sacrificing general capabilities. By clarifying a long-standing misconception and introducing a principled localized tuning strategy, we advance fine-tuning from an underestimated baseline to a leading method for model editing, establishing a solid foundation for future research.
Community
This paper revives fine-tuning for model editing, proving that by simply fixing its flawed implementation and applying a localized parameter update strategy, the simplest method becomes state-of-the-art, handling an unprecedented 100K sequential edits on models up to 72B parameters.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Robust Knowledge Editing via Explicit Reasoning Chains for Distractor-Resilient Multi-Hop QA (2025)
- Data Efficient Adaptation in Large Language Models via Continuous Low-Rank Fine-Tuning (2025)
- Diagnosing Model Editing via Knowledge Spectrum (2025)
- KEDAS: Knowledge Editing Alignment with Diverse Augmentation and Self-adaptive Inference (2025)
- Not All Parameters Are Created Equal: Smart Isolation Boosts Fine-Tuning Performance (2025)
- Continuously Steering LLMs Sensitivity to Contextual Knowledge with Proxy Models (2025)
- A Dual-Axis Taxonomy of Knowledge Editing for LLMs: From Mechanisms to Functions (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper