RELATION EDITING FOR LARGE LANGUAGE MODELS
TL;DR Summary
This study introduces the task of relation editing in large language models, revealing that current methods retain outdated information at rates up to 98.20%. A novel Forgetting-and-Editing framework and a self-paced learning strategy are proposed, significantly improving editing
Abstract
Knowledge editing is a critical technique for the routine updating and maintenance of LLMs. Existing research predominantly assumes changes only to the object within subject-relation-object triples, with minimal exploration into techniques for editing the relation. We term this task Relation Editing (distinct from the established “Object Editing” paradigm). We first construct a dedicated relation editing dataset and benchmark existing algorithms, revealing a critical flaw: even with successful edits, prominent methods suffer from the persistent retention of outdated information, with rates reaching as high as 98.20%. Editing failures stem primarily from two sources: the persistent retention of outdated relationships and the presence of challenging editing samples. To address the first issue, we propose a novel relation editing framework called Forgetting-and-Editing (FE). We theoretically show that existing forgetting methods (i.e., model unlearning) are unsuitable for this purpose and, to this end, introduce a new target assignment strategy within our framework. To mitigate the second challenge, we introduce a self-paced learning strategy, instantiated in a new algorithm named self-paced AlphaEdit (SPaEdit). We conduct extensive experiments on our compiled relation-editing dataset and established object-editing benchmarks. Results demonstrate that our proposed relation editing strategy achieves satisfactory performance on the relation editing task. In addition, SPaEdit outperforms existing SOTA methods on object-editing benchmarks. Our research also suggests further study is warranted in relation editing, particularly on forgetting existing relations.
Mind Map
In-depth Reading
English Analysis
1. Bibliographic Information
1.1. Title
The central topic of this paper is Relation Editing for Large Language Models.
1.2. Authors
The authors are listed as Anonymous authors as the paper is under double-blind review. Therefore, their research backgrounds and affiliations are not disclosed in the paper.
1.3. Journal/Conference
The paper states Paper under double-blind review, indicating it is submitted to a conference, likely ICLR 2025 given the reference format and mentions in the text (e.g., The Thirteenth International Conference on Learning Representations, ICLR, 2025). ICLR is a highly reputable and influential conference in the field of deep learning and artificial intelligence.
1.4. Publication Year
The publication year is not explicitly stated for the paper itself, but it references ICLR, 2025 for some of its own cited works, suggesting it's either published or submitted for publication in 2025.
1.5. Abstract
Knowledge editing is a crucial technique for updating and maintaining large language models (LLMs) without costly full retraining. Existing research primarily focuses on Object Editing, where only the object in a (subject, relation, object) triple is changed. This paper introduces a new task called Relation Editing, which focuses on modifying the relation while keeping the subject and object constant. The authors construct a dedicated relation editing dataset (ReEditBench) and benchmark existing object-editing algorithms, revealing a significant flaw: even if new knowledge is successfully acquired, these methods retain outdated information at very high rates (up to 98.20%).
The paper identifies two main sources of editing failures: the persistent retention of outdated relationships and the presence of challenging editing samples. To address the first issue, a novel Forgetting-and-Editing (FE) framework is proposed. The authors theoretically demonstrate that existing model unlearning strategies are unsuitable for this purpose and introduce a new target assignment strategy within FE. To mitigate the second challenge of difficult samples, a self-paced learning strategy is introduced and instantiated in a new algorithm named Self-paced AlphaEdit (SPaEdit).
Extensive experiments on ReEditBench and established object-editing benchmarks show that the proposed FE strategy significantly improves relation editing performance. Furthermore, SPaEdit outperforms existing state-of-the-art methods on object-editing benchmarks. The research concludes by emphasizing the need for further study in relation editing, especially concerning effectively forgetting existing relations.
1.6. Original Source Link
The original source link is /files/papers/6951e7e69c764da3f20e3720/paper.pdf. This link points to a PDF file, and given the "Anonymous authors Paper under double-blind review" note, it is likely a preprint or submission to a conference.
2. Executive Summary
2.1. Background & Motivation
The core problem this paper aims to solve is the efficient and precise modification of factual knowledge within Large Language Models (LLMs) without undergoing expensive full retraining. LLMs are inherently static once trained, making knowledge editing a critical technique for their routine updating and maintenance, such as correcting inaccuracies or adding new information.
Prior research in knowledge editing has predominantly focused on Object Editing, where only the object (o) in a (subject, relation, object) fact triple is updated (e.g., changing "Paris is the capital of France" to "Paris is the capital of fashion"). However, the paper highlights a significant gap: editing the relation (r) itself, while keeping the subject and object constant, has received minimal attention. Such relation editing tasks are common in practice; for example, changing "Zinedine Zidane is a player for Real Madrid" to "Zinedine Zidane is a coach of Real Madrid" involves updating the relation from player for to coach of. Existing object-editing methods are ill-suited for this, failing to properly erase outdated information and struggling with challenging edits.
The paper's entry point is to formally define this overlooked task as Relation Editing and to address its specific challenges. Their innovative idea is to develop a comprehensive framework that combines explicit forgetting of old relations with self-paced learning for more effective and robust knowledge editing.
2.2. Main Contributions / Findings
The paper makes several primary contributions to the field of knowledge editing for LLMs:
- Formalization of
Relation Editing: The paper formally defines and distinguishesRelation Editingfrom the establishedObject Editingparadigm, highlighting its practical importance and unique challenges. - Construction of
ReEditBench: A dedicatedrelation editing datasetnamedReEditBenchis constructed. This dataset is crucial for benchmarking and evaluating methods specifically designed forrelation editing, filling a gap in existing resources. - Identification of Key Challenges: Through benchmarking existing
object-editing algorithmsonReEditBench, the paper reveals two critical flaws:- Persistent Retention of Outdated Information: Even when successfully learning new relations, existing methods suffer from exceptionally high retention rates (up to
98.20%) of the original, outdated knowledge. This indicates an additive rather than a corrective overwrite. - Failure on Challenging Samples: Existing methods perform poorly on
hard-to-editrelations, where the difference between the model's knowledge of the(subject, new_relation)pair and theobjectis large.
- Persistent Retention of Outdated Information: Even when successfully learning new relations, existing methods suffer from exceptionally high retention rates (up to
- Proposal of
Forgetting-and-Editing (FE)Framework: To address the issue of persistent retention, the paper proposes a novel framework calledForgetting-and-Editing (FE).- It theoretically demonstrates the unsuitability of conventional
model unlearningstrategies (e.g., setting targets to "I don't know" or random responses) forrelation editingdue to systematic biases. - It introduces a new,
interpolation-based target assignment strategywithinFEthat effectively suppresses systematic bias, improves edit success, and reduces retention.
- It theoretically demonstrates the unsuitability of conventional
- Introduction of
Self-paced AlphaEdit (SPaEdit): To mitigate the challenge ofhard editing samples, the paper integratesself-paced learninginto aknowledge editing algorithm, resulting inSPaEdit. This method learns from easier samples first and progressively incorporates more challenging ones, leading to more robust optimization. - Empirical Validation and State-of-the-Art Performance:
-
Experiments show that the
FEstrategy significantly enhances the performance of existingobject-editingmethods onrelation editing tasks, leading to averageSuccessmetric improvements of10.07%and peak improvements of34.49%. -
Combining
FEwithSPaEdityields the bestrelation editingperformance. -
SPaEditalso outperforms existing state-of-the-art methods, includingAlphaEdit, on establishedobject-editing benchmarkslikeZsREandCounterFact, particularly excelling on hard subsets.These findings collectively solve the problem of ineffective
relation editingby providing mechanisms for explicitly forgetting old knowledge and robustly learning new, challenging edits, thereby makingLLMupdates more precise and reliable.
-
3. Prerequisite Knowledge & Related Work
3.1. Foundational Concepts
To understand this paper, a foundational grasp of several core concepts in Large Language Models (LLMs) and machine learning is essential:
- Large Language Models (LLMs): These are advanced
neural networks, typicallyTransformer-based, trained on vast amounts of text data to understand, generate, and process human language. They learn to predict the next word in a sequence, thereby acquiring extensive factual knowledge and reasoning capabilities. Examples includeGPT-J,GPT2-XL, andLLaMA3. - Knowledge Editing: A technique to modify specific factual associations stored within an
LLMwithout the prohibitive cost of retraining the entire model. This is crucial for updating outdated information or correcting errors efficiently. It aims to achieveedit precision(accurately changing the target fact) andknowledge retention(not affecting unrelated facts or general capabilities). - (Subject, Relation, Object) Triples: A common way to represent factual knowledge, often denoted as
(s, r, o). For example,(Paris, is capital of, France).Subject (s): The entity about which a fact is stated.Relation (r): The predicate or relationship linking the subject and object.Object (o): The entity that completes the fact with the subject and relation.
- Object Editing: The traditional
knowledge editingparadigm where theobject (o)in a(s, r, o)triple is changed (e.g., ). - Relation Editing: The novel
knowledge editingparadigm proposed in this paper, where therelation (r)in a(s, r, o)triple is changed (e.g., ). - Parametric vs. Non-Invasive Editing:
- Parametric (Weight-space) Editors: These methods directly modify the
weights(parameters) of theLLMto embed new knowledge. Examples includeROME,MEMIT,AlphaEdit. They aim for surgical alterations. - Non-Invasive Approaches: These methods do not directly alter the
LLM's core weights but use external memory orprompt-based adaptation. Examples includeMELO.
- Parametric (Weight-space) Editors: These methods directly modify the
- Model Unlearning (Machine Unlearning): The process of removing specific data or knowledge from a trained
machine learning modelsuch that the model behaves as if it was never trained on that data. InLLMs, this might involve forgetting specific facts to mitigate bias or comply with privacy regulations. - Self-Paced Learning (SPL) / Curriculum Learning (CL): A training strategy inspired by how humans learn, where models are initially trained on "easy" samples and then gradually exposed to "harder" ones.
SPLautomates this process by dynamically determining sample difficulty. - Linear Regression: A statistical method that models the relationship between a
dependent variableand one or moreindependent variablesby fitting a linear equation to observed data. Inknowledge editing, the process of finding model updates can often be formulated as alinear regression problemwhereinput features(keys) are mapped tooutput targets(values). - Feed-Forward Network (FFN): A component within each
Transformerlayer inLLMsthat processes information independently for each position. Manyknowledge editingmethods target theweightsof theseFFNsas they are thought to store factual knowledge. - Null Space: In
linear algebra, thenull spaceof a matrix is the set of all vectors for which . Inknowledge editing, projecting updates into anull spaceof existing knowledge aims to ensure that the edit does not affect those preserved facts.
3.2. Previous Works
The paper discusses several prior works, mainly categorizing them into Parameter-Based Knowledge Editing, Temporal Adaptation and Unlearning, and Curriculum and Self-Paced Learning.
3.2.1. Parameter-Based Knowledge Editing
These methods directly modify the LLM's parameters. They are split into meta-learning and locate-then-edit approaches.
-
Meta-Learning Approaches:
- : A pioneering work in
knowledge editingthat uses ameta-learningframework to enable models to learn how to update their knowledge. This often involves training a separate model to predict weight changes. MEND (Mitchell et al., 2022): Improvesmeta-learningforknowledge editingby generating low-rank updates that are specific to the edit. It learns a meta-model that predicts a gradient transformation to apply to the network weights.
- : A pioneering work in
-
Locate-then-Edit Approaches: These methods first identify the specific
weightsassociated with a fact and then apply aclosed-form update. They formulateknowledge editingas alinear regression taskwheresubject-relation embeddings(keys) are mapped toobject embeddings(values).- General Formulation:
$
\underset { \Delta } { \operatorname* { m i n } } | ( \mathbf { W } + \Delta ) \mathbf { K } _ { 1 } - \mathbf { V } _ { 1 } | _ { F } ^ { 2 } + \alpha | ( \mathbf { W } + \Delta ) \mathbf { K } _ { 0 } - \mathbf { V } _ { 0 } | _ { F } ^ { 2 } + \beta | \Delta | _ { F } ^ { 2 }
$
Where:
- : The original model's
weight matrix. - : The
perturbation matrixto be learned. - , :
Keysandvaluesfor the new facts to be incorporated. - , :
Keysandvaluesfor existing knowledge to be preserved. - , :
Hyperparametersbalancing theupdate error,preservation error, andregularizationon the perturbation magnitude.
- : The original model's
ROME (Meng et al., 2022):(Rank-One Model Editing)identifies a specificfeed-forward key-value subspacefor a target fact and applies an optimalrank-one perturbationto rewrite it, while minimizingKL-divergenceto preserve general behavior.MEMIT (Meng et al., 2023):(Mass-Editing Memory in a Transformer)scalesROMEto edit thousands of facts simultaneously by applyingrank-one updatesto multipleMLPlayers. It uses aleast-squares objectivewithregularizersforlocality.LoFiT (Yin et al., 2024):(Localized Fine-Tuning)focuses onfine-grained editsat theneuron-levelorhead-level.FiNE (Pan et al., 2025):(Fine-grained Neuron-level Editing)is anotherneuron-level knowledge editing techniqueforLLMs, aiming for precise localization of memories.RECT (Gu et al., 2024):(Regularization to the Rescue)reformulatesmodel editingas alow-rank, layer-wise correction problem. It identifiescausally critical MLP layersand appliesrank-rupdates, using aconsistency lossto preserve original distribution.NSE (Jiang et al., 2024):(Neuron-level Sequential Editing)reframesknowledge editingasneuron-level interventionwithinfeed-forward layers. It usesintegrated-gradients attributionto detect sparse neurons and appliesfact-specific scaling vectorsandadditive bias terms.PRUNE (Ma et al., 2025):(Perturbation-Restrained Sequential Model Editing)treatsmodel editingasparameter-efficient subspace pruningwithinMLPblocks. It identifies asparse maskand trains alow-rank adapteron the pruned subspace.AlphaEdit (Fang et al., 2025): This is a direct predecessor and a key baseline forSPaEdit. It introduces anull-space projection(via matrix ) to theoretically guarantee that updates do not disturb previously stored knowledge. The update is projected into thenull spaceofpreserved knowledge keys(). Its core update rule for is: $ \Delta \mathbf { P } = ( \mathbf { V } _ { 1 } - \mathbf { W K } _ { 1 } ) \mathbf { K } _ { 1 } ^ { \top } \mathbf { P } \left( \mathbf { K } _ { 1 } \mathbf { K } _ { 1 } ^ { \top } \mathbf { P } + \beta \mathbf { K } _ { p } \mathbf { K } _ { p } ^ { \top } \mathbf { P } + \alpha \mathbf { I } \right) ^ { - 1 } $ Where is theidentity matrix, and other symbols are as defined above, with representingpreserved knowledge keys.AlphaEditis known for itssafetyandefficacy.
- General Formulation:
$
\underset { \Delta } { \operatorname* { m i n } } | ( \mathbf { W } + \Delta ) \mathbf { K } _ { 1 } - \mathbf { V } _ { 1 } | _ { F } ^ { 2 } + \alpha | ( \mathbf { W } + \Delta ) \mathbf { K } _ { 0 } - \mathbf { V } _ { 0 } | _ { F } ^ { 2 } + \beta | \Delta | _ { F } ^ { 2 }
$
Where:
3.2.2. Temporal Adaptation and Unlearning
This category focuses on removing or isolating obsolete knowledge.
- Gradient-based Approaches:
Forgetting losses (Yao et al., 2024): Methods that introduce specific loss functions to encourage the model to forget certain data.Orthogonal projection updates (Hoang et al., 2024): Usinggradient projectionto minimizeunlearning interference.Fisher weighted masking (Cha et al., 2024): UtilizesFisher informationto identify importantweightsforunlearning.
- Memory-centric Methods (External Memory):
GRACE (Hartvigsen et al., 2023):(Aging with GRACE)usesdiscrete key-value adaptorsforlifelong model editing.T-Patcher (Huang et al., 2023):(Transformer-Patcher)aims to localize and correct mistakes by modifying a small number of neurons.KV scrubbing (Wang et al., 2024a): Rethinksknowledge memoryforlifelong model editing. These methods often assignfixed forget-set targets(e.g., "I don't know" or random answers), which this paper theoretically shows to inducesystematic biasinlinear regression-based editing.
3.2.3. Curriculum and Self-Paced Learning
Curriculum Learning (CL) (Bengio et al., 2009): Orders training samples from easy to hard, usingheuristics.Self-Paced Learning (SPL) (Kumar et al., 2010): Automates sample selection withregularized weights. These principles have been extended toRL controllers (Graves et al., 2017),LLM instruction-tuning, andcontinual learning (Ke et al., 2022; Liu et al., 2024b; Ge et al., 2025). However, the paper notes that they have not been systematically applied toknowledge editingbefore.
3.3. Technological Evolution
The field of knowledge editing has evolved from initial fine-tuning approaches (which are costly and risk catastrophic forgetting) to more targeted meta-learning and locate-then-edit methods. Early locate-then-edit methods like ROME and MEMIT focused on efficient object editing by identifying and updating specific MLP weights. More recently, work like AlphaEdit introduced null-space projection to guarantee knowledge preservation, addressing concerns about unintended side effects. Concurrently, model unlearning emerged to address the challenge of removing unwanted knowledge.
This paper's work fits within this timeline by addressing a previously overlooked aspect: relation editing. It builds upon the locate-then-edit paradigm, specifically AlphaEdit, by integrating a novel unlearning strategy (FE) and a self-paced learning approach (SPaEdit). This represents an evolution towards more robust, comprehensive, and nuanced knowledge editing capabilities that can handle complex updates and challenging samples while preserving general model integrity.
3.4. Differentiation Analysis
Compared to the main methods in related work, this paper's approach offers several core differences and innovations:
- Focus on
Relation Editing: The most significant differentiation is the explicit focus onRelation Editing. While prior workRaKE (Wei et al., 2023)briefly touched upon it, this paper presents the first systematic study of this task, including a dedicated dataset (ReEditBench) and a tailored methodology. Existingobject-editingmethods are shown to perform poorly onrelation editingdue to the high retention of old knowledge. - Novel
Forgetting-and-Editing (FE)Strategy:- Theoretical Unsuitability of Conventional Unlearning: The paper rigorously shows that standard
model unlearningstrategies (e.g., setting targets to "I don't know" or random responses) introducesystematic biasand are ineffective forrelation forgettinginlinear regression-based editing. - Interpolation-based Target Smoothing:
FEproposes a novelinterpolation-based target assignment strategyfor forgetting. Instead of a fixed or random target, it smoothly moves the old object's representation towards a neutral state, which is theoretically proven to suppresssystematic bias, improveedit success, and reduceretentionwhile inducing smaller perturbations to normal knowledge. This is a critical innovation over existingunlearning methods.
- Theoretical Unsuitability of Conventional Unlearning: The paper rigorously shows that standard
- Integration of
Self-Paced LearningviaSPaEdit:- Addressing
Hard Samples: Unlike existingknowledge editingmethods that typically usesingle-pass optimizationand struggle withdifficult editing samples,SPaEditintegratesself-paced learning(aneasy-to-hard curriculum). This allows the model to build a robust foundation on easier edits before progressively tackling more challenging ones. - Minimal Overhead, Enhanced Robustness:
SPaEditextendsAlphaEditwith minimal structural overhead (only introducing a diagonal matrix ), yet it achieves substantial performance gains, particularly onhard samples, without degradinggeneral capabilities.
- Addressing
- Comprehensive Solution for
Relation Editing: The combination ofFE(for effective forgetting) andSPaEdit(for robust learning, especially of hard samples) provides a holistic framework specifically tailored for the complexities ofrelation editing, where both the erasure of old relations and the acquisition of new ones are critical. This goes beyond the capabilities of existing methods which either focus only on learning new objects or use less effective unlearning mechanisms.
4. Methodology
The paper proposes a novel framework for Relation Editing that addresses the challenges of persistent retention of outdated knowledge and poor performance on hard-to-edit samples. The framework consists of two main components: the Forgetting-and-Editing (FE) strategy, which includes a new target assignment scheme for forgetting old relations, and Self-Paced AlphaEdit (SPaEdit), an algorithm that leverages self-paced learning to handle samples of varying difficulty.
The overall framework is illustrated in Figure 3.

该图像是一个示意图,展示了我们提出的关系编辑框架,结合了遗忘与编辑(FE)策略和自适应AlphaEdit(SPaEdit)算法。图中展示了编辑和遗忘数据的选择过程,以及更新新对象向量的公式 v(ô) = v(o) + eta[v(IDK) - v(o)]。
Figure 3: Overview of our proposed framework for relation editing, combining a novel forgetting-and-editing (FE) strategy with a Self-paced AlphaEdit (SPaEdit) algorithm.
4.1. Problem Description
Knowledge editing aims to update factual triples in LLMs. Unlike object editing, which modifies in a (subject, relation, object) tuple (s, r, o) to , relation editing alters the relation to , resulting in a new tuple .
In the locate-then-edit paradigm, each edit applies a perturbation to the model parameters . Here, and denote the dimensions of the FFN's intermediate and output layers, respectively. For updating relation facts, let be the key matrix for the raw subject-relation pairs, and be the key matrix for the updated subject-relation pairs. The value matrix remains unchanged, representing the object .
Directly applying object editing methods to relation editing would involve minimizing the error for updated relations while preserving existing knowledge. This is typically formulated as:
Where:
-
: The
perturbationto the model weights. -
: The original
weight matrixof theLLM. -
: The
key matrixfor the newsubject-relationpairs . -
: The
value matrixfor theobject. -
: The
squared Frobenius norm, which measures the difference between the model's output and the target values.However, empirical evaluation revealed that this direct application leads to high
retentionof original knowledge and poor performance onhard-to-editrelations, necessitating a new strategy.
4.2. Theoretical Investigation of Forgetting Strategies
The paper theoretically investigates why conventional LLM unlearning methods are unsuitable for old relation forgetting under linear regression-based editing methods like AlphaEdit and MEMIT. These conventional methods typically set the prediction target for data to be forgotten to either "I don't know" (IDK) or a random response.
The analysis models knowledge editing as a linear homogeneous regression problem with a training set . The set is split into (normal data) and (forgetting data). The optimal weight vector is obtained by minimizing Mean Squared Error (MSE):
Where:
- : The optimal
weight vector. - : The
feature matrixfor all data. - : The
label vectorfor all data. - : The contribution from
normal data. - : The contribution from
forgetting data.
4.2.1. Case 1: Fixed Target (e.g., IDK)
If each label in is fixed to a constant (simulating all objects changed to IDK), the solution becomes:
Where:
- : The optimal
weight vectorwhen forgetting data targets a constant value. - : The solution achieved by only applying
normal data. - : The
constant target value(e.g., IDK). - : The total number of samples.
- : The
mean feature vectorof theforgetting data. This equation shows that asystematic biasis introduced, pulling the solution towards . The degree of distortion depends on the correlation between new inputs and .
4.2.2. Case 2: Random Target
If each label in is set to a random value (simulating random object assignments), the expected solution becomes:
\mathbb { E } [ { \pmb w } _ { \mathrm { r a n d } } ^ { * } ] = ( \mathbf { X } ^ { \top } \mathbf { X } ) ^ { - 1 } ( \mathbf { X } _ { g } ^ { \top } { \pmb y } _ { g } + \mathbb { E } [ \mathbf { X } _ { b } ^ { \top } { \pmb y } _ { b } ] ) = { \pmb w } _ { g } ^ { * } + ( \mathbf { X } ^ ^ { \top } \mathbf { X } ) ^ { - 1 } ( 0 . 5 | \mathbb { D } _ { b } | \mathbb { E } [ \mathbf { x } ] ) .
Where:
- : The
expected optimal weight vectorwhen forgetting data targets random values. - : The
expected feature vector. Similar to the first case,random noiseintroduces asystematic biasin expectation, pulling the solution towards a direction determined by theirrelevant feature mean, forcing predicted values to skew towards0.5(average response inLLMs). This distortion affects bothnormalandunlearning samples.
The theoretical analysis concludes that standard unlearning strategies cause normal knowledge to become systematically distorted when used with current model editing methods to forget old relations.
4.3. Knowledge Forgetting via Target Smoothing
Given the ineffectiveness of conventional unlearning strategies, the paper proposes a novel target assignment strategy for knowledge forgetting. The key is to determine a suitable object for the triplet (s, r, o) to be unlearned, such that is neither uniform nor randomly assigned.
The strategy is guided by three considerations:
-
Non-constant assignment: Avoids uniform targets.
-
Non-random assignment: Avoids uncontrolled random targets.
-
Target vector proximity: Ensures the difference between the vector representation of (
v(ô)) and the original object (v(o)) is not too large. This is important because the vector representations of(s, r)and are often highly similar (as shown in Figure 2), and a large disparity in object values would make theoptimization problemsignificantly harder.Based on these, the vector for is generated through
interpolation: Where:
-
: The interpolated
value vectorfor the new targetobjectto be unlearned. -
: The original
value vectorof theobject. -
: The
value vectorrepresenting "I don't know". -
: The
interpolation factor, ahyperparametercontrolling the degree of interpolation. ensures it's a blend.This
assignment strategycreates a non-constant, data-dependentbias termthat suppressessystematic bias, improvesedit success, and reducesretention, while inducing smaller perturbations tonormal knowledgeand yielding more stable optimization.
4.3.1. Theoretical Analysis of the Forgetting-and-Editing Strategy (Appendix B.1)
The paper further analyzes this FE strategy within the linear regression framework. For any sample in the forgetting set , the modified target label (value vector) becomes:
Extending this to a forgetting set of samples, the modified label vector is:
Substituting this into the closed-form solution of the linear regression problem (similar to Eqn. 2) yields:
Where:
- : The optimal
weight vectorwhen using theFE strategy. - : The solution trained solely on
normal data. - : The
original label vectorfor theforgetting set. - : The
IDK label vector(all elements are ). TheFE strategyintroduces a non-constant, data-dependentbias term. This term providesprecise and continuous controloverforgetting strength(via ), preventssystematic bias, preservesprediction diversity, and leads to stableoptimizationdue to proximity between original and target features.
4.4. The Proposed Forgetting-and-Editing (FE) Strategy
The Forgetting-and-Editing (FE) strategy is a comprehensive framework that integrates the unlearning of outdated relations with the injection of new knowledge into a single optimization step. For a batch of relation editing samples, where the -th sample involves changing from to , the procedure involves two stages:
-
Stage 1: Constructing the Forgetting Pairs.
- For each original triplet , the
interpolated target valueis computed using Eqn. 5. - A
forgetting pairis formed, where is thekey vectorcorresponding to theoriginal subject-relation. This pair instructs the model to shift the representation of theold relationtowards aneutral state, effectivelysuppressingtheoutdated knowledge.
- For each original triplet , the
-
Stage 2: Constructing the Editing Pairs.
- Simultaneously, for each new triplet , a
standard editing pairis constructed. Here, is thekey vectorfor thenew subject-relation, and is thetarget valueof theobject. This pair ensures the model accurately captures thenew relational association.
- Simultaneously, for each new triplet , a
-
Joint Optimization.
- Both the
forgetting pairsandediting pairsare concatenated to form thefull training setfor the current batch: - This combined dataset is then fed into a
base editor(e.g.,AlphaEditorSPaEdit). By jointly optimizing for both objectives (forgetting and editing), the algorithm updates the weights to simultaneouslyunlearn the old relationandacquire the new one, resolving the inherent conflict inrelation editing.
- Both the
4.5. Improvement via Self-Paced Learning (SPaEdit)
The FE strategy is further enhanced by incorporating self-paced learning (SPL) to address the challenge of hard editing samples. This approach, named Self-paced AlphaEdit (SPaEdit), builds upon AlphaEdit by introducing an easy-to-hard curriculum.
4.5.1. Formulation of the Multi-Objective Optimization Problem (Appendix B.2)
The fundamental goal of parameter-modifying knowledge editing is to find a minimal perturbation to a model's weight matrix , such that the edited model reflects new knowledge without catastrophically forgetting existing information.
The original AlphaEdit objective for finding an optimal perturbation is:
Where:
-
: The original
weight matrix. -
: The
perturbationto the model. -
: The
null-space projector matrix, which is symmetric () and idempotent (). This matrix projects the update into a subspace that does not interfere with preserved knowledge. -
, :
Keysandvaluesof the facts to be edited (new knowledge). -
:
Regularization coefficientconstraining the overall magnitude of the update . -
:
Regularization coefficientpenalizing interference with previously edited/preserved knowledge represented bykeys.To introduce
self-paced learning,SPaEditrecasts this objective by introducingbinary selectorsto build anadaptive curriculum: Where: -
: A vector of
binary selectorswhere means the -th sample is included in the editing process, and means it is excluded. -
: The
sample-wise lossfor the -th edit, defined as thesquared error: $ \ell _ { i } ( \Delta ) = | ( \mathbf { W } + \Delta \mathbf { P } ) k _ { i } - v _ { i } | _ { 2 } ^ { 2 } = | \Delta \mathbf { P } k _ { i } - r _ { i } | _ { 2 } ^ { 2 } $ Here, is theresidualfor the -th sample (the error the edit needs to correct). -
: The
pace parameter, which controls thedifficulty thresholdof the curriculum. A larger allows more difficult samples to be included. -
The term : This term encourages the inclusion of more samples as increases, balancing the
loss minimizationwith thecurriculum progression.
4.5.2. Alternating Minimization
SPaEdit optimizes this objective via alternating minimization between and .
4.5.2.1. Step 1: Solving for with fixed
With fixed, the problem reduces to a regularized least-squares objective over the subset of "easy" samples (where ). Let be a diagonal matrix with on its diagonal. The objective becomes:
Where represents the residual matrix. Since , . The objective can be rewritten (as derived in Appendix B.3, and similar to Eqn. 31):
$
\operatorname* { m i n } _ { \mathbf { \Delta } } \mathcal { L } ( \mathbf { \Delta } ) = | ( \mathbf { \Delta } \mathbf { P } \mathbf { K } _ { 1 } - \mathbf { R } ) \mathbf { Z } | _ { F } ^ { 2 } + \alpha | \mathbf { \Delta } \mathbf { P } | _ { F } ^ { 2 } + \beta | \mathbf { \Delta } \mathbf { P } \mathbf { K } _ { p } | _ { F } ^ { 2 }
$
This is a convex problem with a closed-form solution for the update :
Where:
- : The effective
projected updatematrix. - : The
residual matrix. - : The
diagonal selection matrixwhere for selected samples. - :
Key matrixfor facts to be edited. - :
Key matrixfor preserved knowledge. - :
Null-space projector matrix. - :
Regularization coefficients. - :
Identity matrix. The matrix is guaranteed to be invertible because and arepositive semi-definite, and (for ) makes the entire matrixpositive definite.
4.5.2.2. Step 2: Determining optimal sample selection with fixed
This step realizes an easy-to-hard curriculum by adjusting the difficulty threshold to progressively incorporate more challenging samples. For each sample :
Where:
-
: The optimal
selectorfor sample given . -
: The
sample-wise lossfor sample with the current . -
: The
pace parameterordifficulty threshold. Samples withlossbelow are considered "easy" and included.This two-step process (
Algorithm 1) is iterated, with gradually increasing over time to incorporate more difficult samples. The iterative process stops when thevalidation loss(calculated on a dedicatedvalidation set)plateaus(early stopping).
Algorithm 1: SPaEdit
Input: (keys for facts to be edited), (values for facts to be edited), (initial model weights), (null-space projector), (keys for preserved knowledge), (regularization coefficients), (initial pace parameter), (pace growth factor), (max iterations).
Output: sequence of edited matrices
-
Initialize: , , , (initially include all samples).
-
Repeat for : a. Update : Calculate
residual. Compute using the closed-form solution (Eqn. 10): $ \Delta _ { \mathrm { S P a E d i t } } ^ { ( t ) } = ( \mathbf { R Z K } _ { 1 } ^ { \top } \mathbf { P } ) ( \mathbf { K } _ { 1 } \mathbf { Z K } _ { 1 } ^ { \top } \mathbf { P } + \beta \mathbf { K } _ { p } \mathbf { K } _ { p } ^ { \top } \mathbf { P } + \alpha \mathbf { I } ) ^ { - 1 } $ Update model weights: . b. Update : Calculatesample-wise lossesfor all samples . Update based on thepace parameter: For each , set if , else . Update . c. Update : . d. Model Selection: Evaluate on thevalidation set(Appendix A.1.2). Ifvalidation lossplateaus(early stopping), break. -
Return: The sequence of edited matrices , with the optimal model selected based on validation.
Compared to
AlphaEdit,SPaEditincurs minimal structural overhead, requiring only the introduction of the diagonal matrix to dynamically control theoptimization orderof the samples.
5. Experimental Setup
5.1. Datasets
The experiments primarily use a newly constructed relation editing dataset and established object editing benchmarks.
5.1.1. ReEditBench (Relation Editing Dataset)
This is a novel benchmark constructed specifically for the Relation Editing task.
-
Source: Curated from high-quality, knowledge-intensive benchmarks, mainly
ZsRE (Levy et al., 2017)andWikidata (Vrandei & Krötzsch, 2014). -
Scale: Total 7,918 high-quality editing instances.
-
Characteristics: Built through a rigorous four-stage pipeline (detailed in Appendix A.1.1 and illustrated in Figure 6):
Knowledge Collection: Sourcing initial(s, r, o)facts fromZsREandWikidata.LLM-based Relation Generation: A generatorLLM() reframes facts intorelation-editing tasksby modifying to while keeping and fixed. Two types of edits are encouraged:New Relation(direct modification, e.g., CEO to CTO) andConditional Relation(adding context/temporal constraint, e.g., President to 46th President).Automated Filtering Pipeline:Script-based Filtering: Checks structural integrity.LLM-based Verification: Uses an independent verifierLLM() to assess factual and semantic plausibility.
Human Validation:30%random sample manually validated,98.5%instances confirmed valid.
-
Examples: Figure 12 provides examples of
ReEditBenchinstances. Each entry is a knowledge replacement task with a subject, relation, original object, and new target object.The following are examples of the
ReEditBenchdataset:
{
"type": "relation",
"step1": {
"subject": "Atlant-Soyuz Airlines",
"src": "What airport is Atlant-Soyuz Airlines associated with?",
"pred": "Vnukovo International Airport",
"rephrase": "Which airport is assigned to Atlant-Soyuz Airlines?",
"alt": "Vnukovo International Airport",
"answers": [ "Sheremetyevo Airport" ],
"loc": "nq question: the polar caps on mars are most probably made up of",
"loc_ans": "water ice",
"cond": "Vnukovo International Airport >> Sheremetyevo Airport || What airport is Atlant-Soyuz Airlines associated with?"
},
"step2": {
"subject": "Atlant-Soyuz Airlines",
"src": "What is the main operational base of Atlan Alliance Airlines?",
"pred": "I don't Know",
"rephrase": "At which airport is Atlant-Soyuz Airlines headquartered, and what serves as its central operational hub?",
"alt": "Vnukovo International Airport",
"answers": [ "Vnukovo International Airport" ],
"loc": "nq question: the polar caps on mars are most probably made up of",
"loc_ans": "water ice",
"cond": "I don't Know >> Vyatka International Airport || What is the main operational base of Atlan Alliance Airlines?"
}
}
In this example, Atlant-Soyuz Airlines changes its associated airport from Vnukovo International Airport to Sheremetyevo Airport. The step1 shows the original fact and its related information, while step2 represents the target edit.
5.1.2. ZsRE (Zero-shot Relation Extraction)
-
Source:
(Levy et al., 2017). -
Characteristics: A benchmark for
zero-shot relation extractiontasks, commonly used forobject editingevaluations. It consists ofsubject-relation-objecttriplets presented as natural language questions and answers. -
Purpose: Used to assess the
universalityandgeneralization capabilitiesofSPaEditon traditionalobject editingtasks. Hard subsets ofZsREare also used for focused evaluation.The following are examples of the
ZsREdataset from Figure 13:
{
"subject": "Shelley's crimsonwing",
"src": "What is the endangered status of Shelley's crimsonwing?",
"pred": "vulnerable",
"rephrase": "What is the conservation status of Shelley's crimsonwing?",
"alt": "vulnerable",
"answers": [ "Endangered" ],
"Lo": "ng question: where is the washington post based out of",
"loc_ans": "Washington, D.C.",
"cond": "vulnerable >> Endangered || What is the endangered status of Shelley's crimsonwing?"
},
{
"subject": "Shelley's crimsonwing",
"src": "What endangered category did the Shelley's crimsonwing finch once fall under?",
"pred": "I don't Know",
"rephrase": "Shelley's' crimson-wing finch was once classified as what level of endangered species?",
"alt": "vulnerable",
"answers": [ "vulnerable" ],
"loc": "ng question: where is the washington post basd ut f",
"loc_ans": "Washington, D.C.",
"cond": "I don't Know >> vulnerable || What is the endangered status of Shelley's crimsonwing?"
}
In these examples, the subject is Shelley's crimsonwing, and the task involves understanding its endangered status, showing factual recall.
5.1.3. CounterFact
-
Source:
(Meng et al., 2022). -
Characteristics: Another benchmark for
object editing, known for its challengingcounterintuitive factual edits. These often require overriding strongpre-existing biasesin the model. -
Purpose: Used to assess
SPaEdit's performance on generative tasks and its robustness against difficult edits. Hard subsets are particularly emphasized.The following are examples of the
CounterFactdataset from Figure 14 (partial JSON):
{
"target_new": "pitcher",
"subject": "Charles Vanel",
"locality_ground_truth": "French"
},
{
"target_new": "Belgium",
"subject": "Nenjil Or Aalayam",
"rephrase_prompt": "Pamukkale's surroundings include",
"locality_ground_truth": "India"
}
These snippets show a subject and a target_new object, often with locality information, aiming to edit specific facts like Charles Vanel's profession or the location associated with Nenjil Or Aalayam.
5.1.4. Validation Set Construction (Appendix A.1.2)
For iterative algorithms like SPaEdit, a dedicated validation set is used for model selection and early stopping.
- Construction:
20%of the full training dataset is randomly held out. For each instance(s, r, o)to , the following are defined:Original Key-Value Pair: for(s, r)and .New Key-Value Pair: for and .Paraphrased Key: (rephrasing of ) forgeneralizationtesting.Forget Target: using theinterpolation factor.
- Iterative Evaluation with Weighted Loss Function: At each iteration ,
perturbationandedited model weightsare obtained. Three distinct losses are calculated:Forgetting Loss(): Measures how successfully the model unlearns the original fact by moving towards . Where denotes the average loss over all samples in the validation set.Efficacy Loss(): Assesses direct acquisition of new knowledge, i.e., error between and .Generalization Loss(): Evaluates applying new knowledge toparaphrased prompts().
- Final Model Selection: The total
validation lossis aweighted sum:Early stoppingwith apatienceof 3 iterations (no improvement within threshold ) is used. The final model is the checkpoint with the best observed . Weights are set ashyperparameters(e.g., ).
5.2. Evaluation Metrics
5.2.1. ReLEditBench Metrics (Relation Editing)
These metrics are designed to holistically evaluate relation editing by assessing both forgetting and learning. For an original fact (s, r, o) and a new fact :
-
Success(): A joint metric verifying if aknowledge editwas successful, requiring two conditions to be met simultaneously: (i) the model must no longer predict the original object for the original query(s, r), and (ii) it must correctly predict the new object for the updated query . Where:- : Expected value over all samples in the dataset .
- :
Indicator function(equals 1 if the condition is true, 0 otherwise). - : The
LLMwith parameters . - : The
probabilityof predictingobjectgivensubject-relation(s, r). - : The
top predictionfor(s, r)is not the original . This means forgetting the old fact. - : The
top predictionfor is the correct new . This means learning the new fact.
-
Retention(): Evaluates whether the model successfully retains the newly introduced knowledge after the edit. It measures theprobabilitythat theold objectis still thetop predictionfor theoriginal prompt(s, r). Note: The formula provided in the paper (Eqn. 21) seems to be forEfficacywith the old relation. Given the context of "Retention (forgetting)", the logical interpretation is the original fact's persistence. For clarity, I will use the common definition of retention in knowledge editing (old fact still predicted). Where:- : Expected value over all samples.
- : The
original objectfor sample . (s, r): Theoriginal subject-relation prompt.- This metric is effectively
old_fact_accuracy, and lower values are better forforgetting.
-
Efficacy(): Measures the model's direct acquisition of the new fact. It is defined as theprobabilitythat thenew objectis thetop predictionfor thenew prompt. A high score signifies successful instantiation of the new knowledge. Where:- : The
target objectfor sample . - : The
new subject-relation prompt.
- : The
-
Generalization(): Evaluates if the model can apply thenew knowledgebeyond the specific prompt it was edited on. It measures the model's ability to predict the correct object when presented with a set ofparaphrasedorsemantically equivalent prompts. Where:- : The
correct object. - : A
set of paraphrased promptsfor thenew subject-relation.
- : The
5.2.2. ZsRE Metrics (Object Editing)
For ZsRE, Efficacy, Generalization, and Specificity are used.
Efficacy(): Same asEfficacyforReEditBench, but forobject editingwhere the relation is fixed and object changes. Measurestop-1 accuracyonedit samplespredicting . Where is the new target object for the -th edit.Generalization(): Same asGeneralizationforReEditBench, measurestop-1 accuracyonequivalent prompts(rephrased statements) for the new knowledge. Where are paraphrased versions of the originalsubject-relationpair for which the model should now predict .Specificity(): Ensureseditingdoes not affectunrelated samples(other facts). Evaluated bytop-1 accuracyof predictions that remain unchanged. Where is thecorrect original objectforunrelated queries. High score means minimal side effects.
5.2.3. CounterFact Metrics (Object Editing)
For CounterFact, Efficacy, Generalization, and Specificity are used (same as ZsRE), plus Fluency and Consistency.
Fluency(,generation entropy): Measures forexcessive repetitionin model outputs. Calculated using theentropyofn-gram distributions(2-gram and 3-gram). Where:- : The
probabilityofbigram. - : The
probabilityoftrigram. Higherentropy(higher score) indicates more diverse andfluent generation.
- : The
Consistency(,reference score): Evaluates theconsistencyof the model's outputs by computing thecosine similaritybetween theTF-IDF vectorsof the model-generated text and areference Wikipedia text. Highercosine similarityindicates betterconsistencywith factual references.
5.3. Base LLMs & Baseline Methods
5.3.1. Base LLMs
Experiments are conducted on three representative LLMs:
- (Meta, 2024)
GPT-J (6B)(Wang & Komatsuzaki, 2021)- (Radford et al., 2019)
5.3.2. Baseline Methods (Appendix A.2)
Seven parametric editing methods are compared:
ROME (Meng et al., 2022):(Rank-One Model Editing)Alocate-then-editmethod that applies arank-one perturbationto aFFNlayer, ensuringlocal rewrite, global preservation.RECT (Gu et al., 2024):(Regularization to the Rescue)Reframesmodel editingas alow-rank, layer-wise correction problemonk contiguous MLP layers, minimizing aconsistency loss.NSE (Jiang et al., 2024):(Neuron-level Sequential Editing)Usesneuron-level interventionviafact-specific scaling vectorsandadditive bias termsonsparse neuron subsets.Fine-Tuning (FT) (Zhu et al., 2020): Formalizesknowledge editingasconstrained fine-tuningof a minimal parameter subset (up- and down-projection matrices of a singleMLP layer) withL2 proximity regularization.MEMIT (Meng et al., 2023):(Mass-Editing Memory in a Transformer)Scalescausal model editingto thousands of facts by applyingrank-one updatesto multipleMLPlayers simultaneously.PRUNE (Ma et al., 2025):(Perturbation-Restrained Sequential Model Editing)Treatsmodel editingasparameter-efficient subspace pruningwithinMLPblocks, training alow-rank adapteron the pruned subspace.AlphaEdit (Fang et al., 2025):(Null-space constrained model editing)Augments thelocate-then-editpipeline with anull-space projectionto prevent updates from disturbing previously stored knowledge, guaranteeingnon-interference.
5.4. Experimental Details (Appendix A.4)
5.4.1. Model Configuration Parameters
The following are the results from Table 4 of the original paper:
| Parameter | Value | Description |
|---|---|---|
model_name |
EleutherAI_gpt-j-6B, gpt2-xl,Llama3-8B | Specifies the pretrained language model. |
layers |
[3-8], [13-17], [4-8] | The target Transformer layers for editing. |
v_num_grad_steps |
25 or 20 | Number of gradient steps for value vector computation. |
vlr |
5e-1 or 1e-1 | Learning rate used during value vector computation. |
v_loss_layer |
27, 47, 31 | The specific model layer used to compute the edit loss. |
kl_factor |
0.0625 | Weight of the KL-divergence regularization term. |
mom2_dataset |
wikipedia | Dataset for computing second-moment statistics. |
rewrite_module_tmp |
Varies by model | Template for the path to the module being rewritten. |
5.4.2. Key Hyperparameters for the SPaEdit and FE Strategies
Forgetting Interpolation Factor (\gamma): For theFE strategy(Eqn. 5). A higher enforces more thorough forgetting.- Set to
0.4forGPT-J-6B. - Set to
0.6for bothLLaMA3-8BandGPT2-XL.
- Set to
Update Regularization Coefficients (\alphaand\beta): For theSPaEdit objective function(Eqn. 7).- : Constrains
overall magnitudeof the update. Set to10. - : Minimizes
impactonpreserved knowledge keys. Set to1.
- : Constrains
Self-Paced Learning Curriculum Parameters (\lambda_0, \mu, T): ForSPaEdit(Algorithm 1).- (Initial Pace Parameter): Initial
difficulty threshold. Set to10. - (Pace Growth Factor): Multiplicative factor for increase per iteration. Set to
1.1. - (Max Iterations): Total number of iterations. Set to
20.
- (Initial Pace Parameter): Initial
6. Results & Analysis
6.1. Core Results Analysis
6.1.1. Results Directly With Object Editing
Initial evaluation of existing object editing methods on the ReEditBench dataset revealed two critical issues, as shown in Figure 1.

该图像是图表,展示了关系编辑中的关键挑战。图(a)比较了不同方法的编辑有效性(蓝色)与原始事实保留率(粉色),显示旧知识的持续存在;图(b)展示了样本难度与编辑成功率之间的强负相关关系,指出在困难样本上性能下降。
Figure 1: Analysis of key challenges in relation editing. (a) The bar chart compares editing efficacy (blue) with Retention of the original fact (pink), showing that old knowledge persists. (b) The scatter plot shows a strong negative correlation between sample difficulty and Efficacy rate, indicating performance decay on challenging samples.
-
Persistent Retention (Figure 1a):
Object editingmethods achieve highsuccess ratesin acquiring new knowledge (blue bars) but concurrently retain the original, conflicting knowledge at exceptionally high rates (pink bars). For instance,AlphaEditonGPT-Jshows asuccess rateof~99%with aretention rateof~98%. This indicates that these methods perform anadditive operationrather than acorrective overwrite, leading to the problematic coexistence ofnewandold knowledge. -
Failure on Hard Samples (Figure 1b): A strong negative correlation exists between the
editing success rateandsample difficulty(measured by themagnitude of the initial residual, ). "Easy samples" (blue) cluster in a high-success region, while "hard samples" (pink) fall into a low-success region. This highlights a consistent failure onhigh-difficulty editing samples.These findings underscore that existing
object editingmethods are ill-suited forrelation editingbecause they fail to eraseoutdated informationand lackefficacyforchallenging edits.
6.1.2. Efficacy of the Forgetting-and-Editing Strategy on Relation Editing
The Forgetting-and-Editing (FE) strategy aims to first forget the old tuple and then incorporate new knowledge. Experiments were conducted using a sequential editing setting with 2000 samples, edited in batches of 100.
The following are the results from Table 2 of the original paper:
| LLMs | Method | Success↑ | Retention↓ | Efficacy↑ | Generalization↑ | ||||
|---|---|---|---|---|---|---|---|---|---|
| Original | +FE | Original | +FE | Original | +FE | Original | +FE | ||
| LLaMA3 | MEMIT | 33.77 | 68.26 (+34.49) | 51.70 | 58.82 (-7.12) | 48.43 | 70.93 (+22.50) | 49.09 | 67.00 (+17.91) |
| RECT | 59.41 | 66.83 (+7.42) | 72.78 | 59.45 (+13.33) | 66.78 | 69.70 (+2.92) | 54.63 | 58.96 (+4.33) | |
| NSE | 43.20 | 54.30 (+11.10) | 53.73 | 52.24 (+1.49) | 45.00 | 58.53 (+13.53) | 59.26 | 58.55 (-0.71) | |
| ROME | 31.39 | 44.91 (+13.52) | 60.47 | 56.36 (+4.11) | 50.91 | 56.64 (+5.73) | 50.93 | 56.80 (+5.87) | |
| FT | 48.88 | 63.45 (+14.57) | 64.49 | 63.57 (+0.92) | 49.96 | 71.01 (+21.05) | 69.16 | 67.31 (-1.85) | |
| PRUNE | 29.40 | 29.81 (+0.41) | 44.68 | 30.46 (+14.22) | 44.04 | 34.25 (-9.79) | 43.86 | 42.97 (-0.89) | |
| AlphaEdit | 52.18 | 78.46 (+26.28) | 78.34 | 67.12 (+11.22) | 79.17 | 83.24 (+4.07) | 76.62 | 80.03 (+3.41) | |
| SPaEdit(Ours) | 54.45 | 81.71 (+27.26) | 68.56 | 62.77 (+5.79) | 83.23 | 87.37 (+4.14) | 75.88 | 81.14 (+5.26) | |
| GPT2-XL | MEMIT | 56.31 | 57.79 (+1.48) | 80.26 | 57.21 (+23.05) | 85.23 | 84.67 (-0.56) | 80.68 | 85.21 (+4.51) |
| RECT | 54.60 | 54.72 (+0.12) | 78.10 | 61.62 (+16.48) | 82.35 | 84.08 (+1.73) | 78.37 | 77.12 (-1.25) | |
| NSE | 45.00 | 45.45 (+0.45) | 58.53 | 58.24 (+0.29) | 59.26 | 59.99 (+0.73) | 58.55 | 59.43 (+0.88) | |
| ROME | 45.74 | 45.82 (+0.08) | 61.71 | 61.49 (+0.22) | 61.70 | 61.39 (-0.31) | 61.19 | 61.78 (+0.59) | |
| FT | 49.96 | 51.32 (+1.36) | 71.01 | 67.25 (+3.76) | 69.16 | 69.93 (+0.77) | 67.31 | 67.58 (+0.27) | |
| PRUNE | 37.88 | 38.04 (+0.16) | 52.62 | 39.14 (+13.48) | 54.49 | 55.71 (+1.22) | 52.99 | 52.60 (-0.39) | |
| AlphaEdit | 65.31 | 75.93 (+10.62) | 91.31 | 50.46 (+40.85) | 86.83 | 87.36 (+0.53) | 84.51 | 85.50 (+0.99) | |
| SPaEdit(Ours) | 62.00 | 83.93 (+21.93) | 68.55 | 48.78 (+19.77) | 85.93 | 88.46 (+2.53) | 87.36 | 87.50 (+0.14) | |
| GPT-J | MEMIT | 72.55 | 82.36 (+9.81) | 92.98 | 77.63 (+5.09) | 82.12 | 82.42 (+0.30) | 84.69 | 82.10 (+0.20) |
| RECT | 72.54 | 77.63 (+5.09) | 91.67 | 74.54 (+17.13) | 81.90 | 82.42 (+0.30) | 84.89 | 82.10 (+0.20) | |
| NSE | 45.65 | 45.95 (+0.30) | 62.13 | 61.12 (+1.01) | 62.03 | 60.94 (-1.09) | 61.52 | 61.63 (+0.11) | |
| ROME | 46.38 | 47.79 (+1.41) | 63.34 | 29.27 (+34.07) | 63.32 | 61.49 (-1.83) | 63.24 | 63.78 (+0.54) | |
| FT | 51.19 | 61.10 (+9.91) | 66.24 | 43.50 (+22.74) | 70.79 | 78.72 (+7.97) | 67.31 | 68.67 (+1.34) | |
| PRUNE | 55.71 | 63.05 (+7.34) | 79.12 | 59.87 (+19.25) | 77.25 | 77.00 (-0.25) | 75.41 | 76.62 (-1.21) | |
| AlphaEdit | 65.99 | 89.98 (+23.99) | 98.20 | 63.84 (+34.36) | 85.53 | 85.64 (+0.11) | 86.87 | 87.80 (+0.93) | |
| SPaEdit(Ours) | 78.46 | 91.02 (+12.56) | 88.24 | 59.84 (+28.40) | 75.93 | 88.08 (+12.15) | 87.36 | 88.58 (+1.22) | |
Analysis:
- Significant Improvement with
FE: TheFEstrategy consistently improves performance across all methods andLLMs. It achieves up to a34.49%increase inSuccess(MEMIT on LLaMA3) and a remarkable40.85%reduction inRetention(AlphaEdit on GPT2-XL). This indicates thatFEgenuinely helps in replacing old knowledge, not just adding new. - SPaEdit's Superiority: When combined with
FE,SPaEditconsistently yields the bestrelation editingperformance. For instance, achieves81.71% SuccessonLLaMA3,83.93%onGPT2-XL, and91.02%onGPT-J, often with lowerRetentionrates compared to otherFE-enhancedmethods. - Addressing Misleading Baselines: The high
Retentionof somebaselineswithoutFE(e.g.,98.20%for AlphaEdit on GPT-J) is misleading, as it stems from lowediting successthat doesn't challenge the original knowledge.FEachieves highediting successwhile effectively forgetting outdated facts. - Residual Retention: Despite improvements,
Retentionremains non-trivial (often around50%in difficult settings), suggesting that completelyclean forgettingis an ongoing challenge.
6.1.3. Analysis of the Forgetting Strategy
Figure 4 presents an empirical comparison of four unlearning strategies.

该图像是比较不同方法在关系编辑任务中的表现图,包括LLama3、GPT2-XL和GPT-J。上方的柱状图展示了各种方法的成功率(Suc)和保持率(Ret),而下方的曲线图则展示了在不同 α 值下的成功率和保持率的变化趋势。
Figure 4: The figure shows two rows of charts. The top row compares the success rate (Suc) and retention rate (Ret) of different forgetting strategies (No-Forget, Forget-IDK, Forget-RND, Ours) across three LLMs (LLaMA3, GPT2-XL, GPT-J). The bottom row displays sensitivity analysis for the interpolation factor lambda (λ), showing how success and retention rates change as λ varies, for AlphaEdit on GPT-J.
Analysis of Unlearning Strategies (Top Row of Figure 4):
- The results clearly validate the theoretical analysis from Section 3.1.
Conventional unlearning strategies(Forget-IDKandForget-RND) which set targets to "I don't know" or random values, are largely ineffective at reducingknowledge retention. For instance, onGPT-J, these approaches result inretention ratesas high as77.2%and77.9%respectively. This confirms that theirinherent systematic biasesimpedeeffective forgetting.- The proposed
FE strategy(labeledOurs), which interpolates thevalue vectorof the outdated fact towards a neutral state, performs exceptionally well. It achieves the best trade-off betweensuccessandretention ratesacross all testedLLMs(LLaMA3,GPT2-XL, andGPT-J). It consistently achieves the lowestRetention ratewhile maintaining highSuccess rates.
Sensitivity Analysis on Hyperparameter (Interpolation Factor) (Bottom Row of Figure 4):
- The sensitivity analysis for the
interpolation factor(denoted aslambdain the figure) reveals a clear trade-off betweenforgettingandlearning. - A larger leads to more effective
forgetting(monotonic decrease inRetention rate). - However, the
Success rateshows aconcave trajectory, increasing initially and then decreasing with higher . - An
optimal windowfor is identified, where theSuccess rateis maximized without significant compromise inforgetting. This wide effective range highlights therobustnessof theFE strategytohyperparameter tuning.
6.1.4. Generalization and Performance on Object Editing Benchmarks
To assess SPaEdit's universality and generalization, it was applied to object-editing benchmarks, focusing on hard subsets of ZsRE and CounterFact (100 examples each).
The following are the results from Table 3 of the original paper:
| LLM | Method | Efficacy↑ | Generalization↑ | Specificity↑ |
|---|---|---|---|---|
| LLaMA3 | ROME | 31.87 | 32.4 | 32.26 |
| MEMIT | 86.07 | 82.39 | 33.33 | |
| AlphaEdit | 81.87 | 78.11 | 33.03 | |
| SPaEdit | 92.32 | 82.6 | 32.11 | |
| GPT2-XL | ROME | 15.87 | 16.98 | 7.74 |
| MEMIT | 71.47 | 63.14 | 7.37 | |
| AlphaEdit | 92.17 | 82.68 | 7.72 | |
| SPaEdit | 98.96 | 89.89 | 7.23 | |
| GPT-J | ROME | 23.69 | 27.9 | 24.12 |
| MEMIT | 94.86 | 90.02 | 28.22 | |
| AlphaEdit | 96.26 | 90.46 | 28.15 | |
| SPaEdit | 99.97 | 91.3 | 28.61 |
Results on ZsRE (Table 3):
SPaEditconsistently establishes a newstate-of-the-arton theZsRE benchmarkacross all tested models.- Its lead in
Efficacyis particularly notable:92.32%onLLaMA3(significantly overAlphaEdit's81.87%) and a near-perfect99.97%onGPT-J. - It also achieves the top score in
Generalization(89.89%onGPT2-XL) and leads inSpecificityonGPT-J(28.61%). - The
hard sample subsetposes a considerable challenge, causing performance degradation for strong methods likeAlphaEdit.SPaEditexcels by maintaining superior performance due to itsstrategic, staged learning process, which avoidsoptimization pitfallsof resolving high-residual errors simultaneously.
Results on CounterFact Hard Subset (Appendix C.2, Table 6): The following are the results from Table 6 of the original paper:
| LLM | Method | Efficacy↑ | Generalization↑ | Specificity↑ | Fluency↑ | Consistency↑ |
|---|---|---|---|---|---|---|
| LLaMA3 | ROME | 32.02 | 33.41 | 34.31 | 425.55 | 13.01 |
| MEMIT | 69.22 | 65.61 | 30.54 | 629.68 | 53.15 | |
| AlphaEdit | 79.21 | 73.54 | 30.92 | 629.91 | 56.67 | |
| SPaEdit (Ours) | 92.80 | 95.21 | 42.51 | 631.11 | 56.78 | |
| GPT2-XL | ROME | 39.42 | 30.01 | 5.82 | 592.64 | 65.09 |
| MEMIT | 70.45 | 72.98 | 7.93 | 465.78 | 53.58 | |
| AlphaEdit | 83.22 | 83.91 | 8.54 | 621.76 | 55.62 | |
| SPaEdit (Ours) | 92.66 | 94.82 | 9.62 | 629.26 | 54.52 | |
| GPT-J | ROME | 32.05 | 37.01 | 25.76 | 514.82 | 15.64 |
| MEMIT | 79.22 | 78.27 | 27.58 | 618.93 | 57.84 | |
| AlphaEdit | 87.52 | 86.13 | 28.76 | 621.80 | 59.28 | |
| SPaEdit (Ours) | 92.77 | 93.12 | 38.73 | 622.52 | 59.66 |
SPaEditachievesnear-perfect Efficacyacross all models on theCounterFact hard subset.- It sets a new
state-of-the-artinFluency(e.g.,631.11onLLaMA3), indicating higher-quality, more natural language generation post-edit. - This is achieved while maintaining strong
GeneralizationandSpecificity, demonstrating arobust and well-balanced editing profile.
Analysis of Sample Difficulty Distribution (Appendix C.2, Figure 7):

该图像是难度分布图,展示了两个数据集的困难样本数量分布情况。左侧为 ZsRE 硬子集,右侧为 CounterFact 硬子集,均表示在难度范围为 0 到 20 中的样本计数。
Figure 7: The image presents two difficulty distribution charts for selected hard subsets. (a) The ZsRE hard subset has a varied difficulty distribution. (b) The CounterFact hard subset is heavily concentrated in the high-difficulty region.
- The evaluation focuses on
curated hard subsetsbecausefull benchmarksare often dominated by simple samples. ZsRE hard subset(Figure 7a) shows amixed difficulty distribution.CounterFact hard subset(Figure 7b) ismore extreme, with nearly all samples concentrated in thehigh-difficulty range, serving as astress test.- This challenge-focused evaluation highlights
SPaEdit's advantage: itsself-paced, easy-to-hard curriculumcan intelligently identify easier samples even within a difficult set to start optimization, leading to robust updates for challenging edits whereone-shot methodsfail.
Full Benchmark Performance and Saturation Analysis (Appendix C.3, Table 7): The following are the results from Table 7 of the original paper:
| LLM | Method | CounterFact | ZsRE | ||||||
|---|---|---|---|---|---|---|---|---|---|
| Eff.↑ | Gen. ↑ | Spe. ↑ | Flu. ↑ | Consis. ↑ | Eff. ↑ | Gen. ↑ | Spe. ↑ | ||
| LLaMA3 | ROME | 64.40 | 61.42 | 49.44 | 449.06 | 3.31 | 2.01 | 1.80 | 0.69 |
| MEMIT | 65.65 | 64.65 | 51.56 | 437.43 | 6.58 | 34.62 | 31.28 | 18.49 | |
| AlphaEdit | 98.90 | 94.22 | 67.88 | 622.49 | 32.40 | 94.47 | 91.13 | 32.55 | |
| SPaEdit (Ours) | 99.24 | 94.62 | 69.37 | 624.69 | 33.73 | 95.72 | 93.07 | 33.25 | |
| GPT2-XL | ROME | 54.60 | 51.18 | 52.68 | 366.13 | 0.72 | 47.50 | 43.56 | 14.27 |
| MEMIT | 94.70 | 85.82 | 60.50 | 477.26 | 22.72 | 79.17 | 71.44 | 26.42 | |
| AlphaEdit | 99.50 | 93.95 | 66.39 | 597.88 | 39.38 | 94.81 | 86.11 | 25.88 | |
| SPaEdit (Ours) | 99.65 | 94.78 | 67.83 | 599.52 | 40.23 | 95.92 | 87.63 | 27.25 | |
| GPT-J | ROME | 57.50 | 54.20 | 52.05 | 589.42 | 3.22 | 56.42 | 54.65 | 9.86 |
| MEMIT | 98.55 | 95.50 | 63.64 | 546.28 | 34.89 | 94.91 | 90.22 | 30.39 | |
| AlphaEdit | 99.75 | 96.38 | 75.48 | 618.50 | 42.08 | 99.79 | 96.00 | 28.29 | |
| SPaEdit (Ours) | 99.82 | 96.82 | 76.23 | 620.35 | 44.33 | 99.83 | 97.12 | 30.47 | |
- Existing
state-of-the-art methodshave achievednear-saturation performanceon the "easy" and "medium" portions of theCounterFactandZsRE datasets. The primaryfailure modefor current technology is in the "hard" tail. SPaEditnot onlydominates on the hard subsetsbut also consistently achieves thebest performanceacross thefull benchmarks, demonstratingrobustnesswhere it matters most.
6.2. Mechanistic Insight Into SPaEdit
Figure 5 provides insight into SPaEdit's internal curriculum dynamics and cost-benefit profile.

该图像是图表,展示了两部分内容。左侧(a)显示了基于难度的自适应学习动态,随时间的推移(T=1至T=13)在不同难度上的表现变化。右侧(b)则展示了成本效益分析,比较了不同算法在处理困难样本时的执行时间和编辑成功率,突出自适应算法(SPaEdit)在保证鲁棒性上的高额成本及其他算法的弱点。
Figure 5: (a) shows easy-to-hard self-paced curriculum dynamics. (b) shows the costbenefit tradeoff: modest extra time yields large efficacy gains on hard samples.
- Curriculum Dynamics (Figure 5a): The plot traces how the
sample-difficulty distributionevolves underself-paced learning.- At the start (), the distribution is
right-skewed(manyhard samples). - As training progresses (),
proficiency increases, and the mass shifts from thehard(right) to theeasy(left) region. - By later iterations (), the distribution is
left-skewed, meaning most samples are easy. This progression demonstrates theeffectiveness of the parameter updates.
- At the start (), the distribution is
- Cost-Benefit Analysis (Figure 5b):
- On tasks with a low proportion of
hard samples,SPaEditincursnegligible overhead, matchingbaselineswhile achievingsuperior efficacy. - As
task difficulty increases,SPaEditstrategically investsmodest additional computation time, yielding asubstantial gaininediting successcompared tobaselineswhose performance degrades sharply. This favorable trade-off demonstratesSPaEdit'sefficient resource allocationandrobustness.
- On tasks with a low proportion of
6.3. Ablation Studies / Parameter Analysis
6.3.1. Comprehensive Ablation Study on Forgetting Strategies (Appendix C.1)
The following are the results from Table 5 of the original paper:
| LLM | Method | No-Forgetting | + FE (IDK) | + FE (Random) | + FE (Ours) | ||||
|---|---|---|---|---|---|---|---|---|---|
| Retention ↓ | Efficacy ↑ | Retention↓ | Efficacy↑ | Retention↓ | Efficacy↑ | Retention↓ | Efficacy↑ | ||
| LLaMA3 | AlphaEdit | 88.34 | 89.17 | 76.11 | 75.23 | 76.90 | 78.19 | 74.50 | 83.24 |
| SPaEdit | 88.56 | 83.23 | 75.92 | 83.48 | 70.41 | 82.17 | 68.56 | 87.37 | |
| GPT2-XL | AlphaEdit | 91.31 | 88.83 | 60.25 | 83.45 | 65.81 | 84.90 | 50.46 | 87.36 |
| SPaEdit | 68.55 | 85.93 | 55.18 | 80.15 | 61.33 | 81.82 | 48.78 | 88.46 | |
| GPT-J | AlphaEdit | 98.20 | 99.53 | 81.67 | 89.12 | 85.43 | 81.30 | 77.84 | 85.64 |
| SSPaEdit | 88.24 | 85.93 | 65.40 | 88.31 | 72.88 | 89.04 | 59.84 | 88.08 | |
Analysis:
- Naive Strategies (IDK, Random) Unfavorable Trade-off: These strategies lower
Retentioncompared toNo-Forgettingbut often at a cost ofEfficacy degradation. ForSPaEditonGPT2-XL,Efficacydrops from85.93%to80.15%withIDK. This shows a difficult trade-off betweenforgetting old factsandlearning new ones. - Our
FEStrategy is Most Effective at Unlearning: The proposedFE strategy(+ FE (Ours)) consistently achieves thelowest Retention rateacross every model and for bothAlphaEditandSPaEdit. For example,SPaEdit'sRetentiononGPT2-XLis reduced to48.78%, proving its state-of-the-art capability in erasing outdated knowledge. - Synergistic Effect: Our
FE strategycreates asynergistic effect, achieving the bestunlearning(lowestRetention) while simultaneously maintaining or significantlyimproving Efficacy. ForSPaEdit, the "Ours" strategy boostedEfficacy(e.g.,83.23%to87.37%onLLaMA3) while achieving the lowestRetentionscores. This confirms that the carefully designedforgetting targetsfacilitate a cleaner, more effective integration of new knowledge.
6.3.2. Impact of Semantic Similarity on Relation Editing (Appendix C.5)
Figure 9 visually investigates the influence of semantic properties on relation editing.

该图像是图表,展示了语义相似性对编辑成功率的影响。部分 (a) 表示不同相似性下的编辑成功率与遗忘成功率,依次为低相似性 (45.2%)、中相似性 (65.8%) 和高相似性 (95.1%)。部分 (b) 描述了语义相似性与编辑成功率之间的关系,散点图显示出两者之间的低相关性,其皮尔逊相关系数约为 0.3。
Figure 9: Analysis of Semantic Similarity. (a) Asymmetric Impact: Semantic proximity facilitates new knowledge acquisition (blue bars rise) but hinders the forgetting of old knowledge (red bars fall, revealing a trade-off. (b) Weak Correlation with Editing Success: The scatter plot reveals high variance between semantic similarity and editing success rates. The weak correlation (Pearson indicates that semantic similarity acts as a noisy predictor, failing to capture the full complexity of editing difficulty compared to the robust signal provided by computational residuals.
- Asymmetric Impact of Semantic Similarity (Figure 9a):
Editing Success(Blue Bars): Shows a strongpositive correlationwithsemantic similarity. As relations becomesemantically closer(e.g., CEO to CTO), theediting success raterises sharply from45.2%to95.1%. This suggests models leverage existing semantic structures.Forgetting Success(Red Bars): Exhibits a clearnegative trend.Forgettingis significantly harder forsemantically close relations(30.7%) compared todistant ones(65.8%). Highsemantic proximitycausesstrong interference, making it difficult to disentangle old from new knowledge.
- Justification for Computational Residual (Figure 9b):
- The scatter plot reveals
semantic similarityis anoisy predictorof performance, withhigh dispersionand aweak correlation(Pearson ) toediting success. - This contrasts with the
computational residual(Figure 1b), which shows a strong, distinctnegative correlationwithsuccess. - The
computational residualacts as aholistic proxyaggregating alllatent influencing factors(semantics, knowledge frequency, structural complexity), providing adirect, quantifiable signalof the actualoptimization barrier. Thus, it is a morerobustandcomputationally efficient standardfor thecurriculum learningthansemantic metrics alone.
- The scatter plot reveals
6.4. General Capability Tests (Appendix C.4)
To evaluate the long-term impact on general capabilities, a sequential editing experiment was conducted on LLaMA3-8B, evaluating performance on six downstream tasks (SST, MRPC, CoLA, RTE, MMLU, NLI) after each batch of edits.
The following are the results from Figure 8 of the original paper:

该图像是图表,展示了不同编辑方法在多个评估任务上的F1分数变化。图中显示了多个方法的表现,包括AlphaEdit、RECT、PRUNE、MEMIT和SPaEdit,横轴为编辑项目数量,纵轴为F1得分。
Figure 8: A comparison of the impact of different editing methods on general capability during sequential editing. Both SPaEdit and AlphaEdit demonstrate exceptional stability, proving the safety of the projection mechanism. The identical stability of SPaEdit confirms that its iterative process does not harm the model's general knowledge.
Analysis:
- Catastrophic Forgetting in Unconstrained Methods:
MEMIT,RECT, andPRUNEshowsevere performance collapse, confirming thatunconstrained, cumulative editsdamage general abilities. - Stability of Single-Step Projection (
AlphaEdit):AlphaEdit'sperformance curveremains almost perfectlyflat, demonstrating thatconstraining edits to a specific subspaceis highly effective atpreserving general capabilities. - SPaEdit's Stability:
SPaEdit'sperformance curveis virtuallyidenticaltoAlphaEdit, providing strong evidence that itsiterative optimization processdoes not degradegeneral capabilities. Each step withinSPaEdit'sself-paced curriculumremains safely within theconstrained subspace, finding a moreprecise solutionfor target knowledge without harmful side effects.
6.5. Robustness Analysis Against Superficial Editing Attacks (Appendix C.6)
To assess SPaEdit's robustness beyond standard metrics, it was evaluated against superficial editing attacks (Xie et al., 2025), which use contextual triggers (Wiki, Rep, Que) to elicit original (pre-edit) knowledge.
The following are the results from Table 8 of the original paper:
| Method | Wiki Attack | Rep Attack | Que Attack | |||
|---|---|---|---|---|---|---|
| OM ↓ | OP ↓ | OM ↓ | OP ↓ | OM ↓ | OP ↓ | |
| ROME | 54.95 | 58.24 | 61.74 | 64.02 | 38.37 | 38.37 |
| MEMIT | 52.75 | 54.95 | 40.15 | 42.42 | 37.21 | 37.21 |
| PMET | 70.33 | 72.43 | 66.67 | 71.97 | 39.29 | 41.67 |
| r-ROME | 54.95 | 57.14 | 64.39 | 68.18 | 40.48 | 40.48 |
| AlphaEdit | 72.53 | 73.62 | 68.18 | 71.97 | 34.52 | 35.71 |
| SPaEdit+FE(Ours) | 50.81 | 27.23 | 38.52 | 33.84 | 33.19 | 35.11 |
Metrics:
Original Match (OM)(): Percentage of times the model's output matches the original (pre-edit) answer. Lower is better.Original Probability (OP)(): Percentage of times the model assigns a higher probability to the original answer than the new answer. Lower is better.
Analysis:
Superficial editingis a significant challenge, with high-performing editors likeAlphaEditshowing considerable vulnerability (over70% OMonWiki attack).SPaEdit+FE (Ours)demonstratesmarkedly superior robustnessacross all threeattack types, achieving the lowest scores for bothOMandOP. For example, on theWiki attack,SPaEditreducesOMto50.81%, a substantial improvement overAlphaEdit(72.53%) andMEMIT(52.75%).- This enhanced
robustnessis attributed to thesynergistic interplayof theFE strategy(actively unlearning outdated tuples) and theself-paced curriculum(encouraging deeper integration of new knowledge).
6.6. Stability Analysis (Appendix C.7)
A stability analysis was conducted by repeatedly (100 times) sampling 100 instances from ZsRE and applying SPaEdit and baselines.
The following are the results from Figure 10 of the original paper:

该图像是一个箱线图,展示了在不同模型(LLama3、GPT2-XL、GPT-J)下四种编辑方法(SPaEdit、AlphaEdit、MEMIT、ROME)的编辑成功率分布。箱线图显示,SPaEdit在各个模型中均表现出更高的成功率和更低的方差,表明其优越的鲁棒性。
Figure 10: Edit stability analysis on the ZsRE benchmark. The box plot illustrates the distribution of editing success rates over 100 trials, each with 100 randomly sampled edits. SPaEdit demonstrates significantly lower variance and a higher median performance compared to baseline methods, indicating superior robustness.
Analysis:
SPaEditconsistently achieveshigh performance(85%to95%) withremarkably low variance, indicatinghigh reliabilityand independence from specific edit samples.AlphaEditshows widervariance(75%to90%).MEMITis more varied (60%to95%).ROMEdemonstrates theleast stability(10%to40%), highly sensitive to chosen instances.- This confirms
SPaEdit'srobustnessandpredictable, consistently high-quality results.
6.7. Iterative Runtime of SPaEdit (Appendix C.8)
The computational cost per iteration of SPaEdit is dictated by the closed-form update for the perturbation matrix , particularly the matrix inversion step (Eqn. 10). The selection matrix makes this sparsity-dependent.
The following are the results from Figure 11 of the original paper:

该图像是图表,展示了不同模型(LLaMA3、GPT-J 和 GPT2-XL)在每次迭代中所需的执行时间。随着迭代次数的增加,LLaMA3的执行时间显著减少,而GPT-J和GPT2-XL的时间相对稳定,表明了自适应学习策略的影响。
Figure 11: SPaEdit iteration time analysis. The plot shows the wall-clock time required for each successive iteration. As the self-paced curriculum incorporates more challenging samples, the computational complexity and thus the execution time per step gradually increase, aligning with our theoretical analysis.
Analysis:
- In initial iterations, is small, is sparse (few "easy" samples), leading to
low computational cost. - As training progresses, increases, becomes denser (more "challenging samples" included), and the
computational complexity(andexecution time per iteration) gradually increases. - Figure 11 empirically confirms this: the
wall-clock timeper iterationgradually increasesas theself-paced curriculumincorporates more challenging samples. This is a deliberate design, allocating more resources only as needed.
6.8. Qualitative Analysis (Appendix C.9)
Qualitative case studies (from Appendix C.9.1, C.9.2, C.9.3) demonstrate the ability to completely and cleanly replace existing factual knowledge.
Case 1: GPT-J editing "Tonel is affiliated with S.C. Beira-Mar" to "Tonel is affiliated with SV Neulengbach".
No-Forget (OE): Generates the outdated "S.C. Beira-Mar".Forget-IDK: Outputs "I don't know. Tonel is often linked to S.C. Beira-Mar.", still referencing the old fact.Forget-RND: Generates "Tonel is affiliated with River... and S.C. Beira-Mar... Madrid.", showing corrupted output with old fact.Ours: Confidently generates "Tonel is affiliated with the Brazilian football club SV Neulengbach.", showing clean replacement.
Case 2: LLaMA3 editing "Publisher of Player's Guide to the Dragonlance Campaign was TSR" to "was Bandai".
No-Forget (OE): Outputs the outdated "TSR".Forget-IDK: Generates "was I don't TSR.", corrupted with the old fact.Forget-RND: Outputs "was TSR. x y z.", corrupted with the old fact.Ours: Generates "The publisher... was Bandai, Bandai, Bandai, Bandai.", showing clean replacement (with some repetition, common for generative models).
Case 3: GPT-J editing "Krassimira Stoyanova has a soprano voice" to "a mezzo-soprano voice".
-
No-Forget (OE): Outputs the outdated "soprano voice". -
Forget-IDK: Generates "voice type is I don't know a soprano.", corrupted with the old fact. -
Forget-RND: Outputs "a soprano. Krassimira Stoyanova's voice type is soprano", failing to forget. -
Ours: Generates "mezzo-soprano a mezzo-soprano. Krassimira Stoyanova has a mezzo-soprano voice.", showing clean replacement.Conclusion:
No-Forgetmethods consistently fail, andconventional unlearning strategiesalso fail by generating corrupted outputs that still reference the forbidden original answer.Our methodis the only one that completely replaces the outdated information with the new target knowledge, achievinggenuine and effective knowledge update.
7. Conclusion & Reflections
7.1. Conclusion Summary
This paper formally defines Relation Editing as a novel and critical task for updating Large Language Models (LLMs), distinguishing it from the traditional Object Editing paradigm. Through extensive benchmarking, the authors identify two major weaknesses of existing knowledge editing methods when applied to relation editing: the persistent retention of outdated information and poor performance on difficult editing samples.
To address these challenges, the paper introduces two key contributions:
-
Forgetting-and-Editing (FE)Framework: This novel framework incorporates a theoretically groundedunlearning strategythat moves away from conventional fixed or random targets. Instead, it uses aninterpolation-based target assignmentto effectively suppresssystematic bias, improveedit success, and reduceretentionof old knowledge. -
Self-paced AlphaEdit (SPaEdit)Algorithm: This algorithm integratesself-paced learning(aneasy-to-hard curriculum) into theAlphaEditframework.SPaEditsystematically learns from easier samples first, gradually incorporating more challenging ones, leading to robust optimization for difficult edits.Extensive experiments on the newly compiled
ReEditBenchdataset confirm that theFEstrategy significantly enhancesrelation editingperformance by enabling effective forgetting. Furthermore,SPaEditnot only excels onrelation editingtasks but also establishes newstate-of-the-artperformance onobject-editing benchmarkslikeZsREandCounterFact, particularly forhard samples. The research also demonstrates thatSPaEditmaintainsgeneral capabilitiesand exhibits superiorrobustnessagainstsuperficial editing attacks.
7.2. Limitations & Future Work
The authors acknowledge that despite the significant gains achieved, Retention remains non-trivial in absolute terms, often around 50% in difficult settings. This indicates that fully clean and permanent forgetting of obsolete relations is still an unsolved problem. Therefore, future work is explicitly suggested to develop more effective unlearning mechanisms specifically tailored for relation editing.
7.3. Personal Insights & Critique
This paper makes a highly valuable contribution by formalizing Relation Editing, a practically relevant but previously overlooked aspect of knowledge editing. The distinction between object and relation editing is crucial, as the paper convincingly demonstrates that existing methods designed for the former fail spectacularly on the latter, primarily due to the inability to properly forget old information.
The FE framework with its interpolation-based target smoothing is an elegant solution to the unlearning problem. The theoretical analysis of why IDK or random targets lead to systematic bias is rigorous and provides a strong foundation for their proposed method. This insight into unlearning within linear regression-based editing is particularly impactful and could be highly transferable to other unlearning contexts where parametric updates are used.
The integration of self-paced learning into AlphaEdit (SPaEdit) is also a smart move. The empirical evidence clearly shows its superiority, especially on hard samples, which are the real test of any editing algorithm. The cost-benefit analysis demonstrating modest overhead for significant efficacy gains is compelling. The stability analysis and robustness against superficial editing attacks further solidify SPaEdit's practical value, as these are critical concerns for LLM deployment.
Potential Issues/Areas for Improvement:
- Absolute Retention Rates: While
FEsignificantly reducesretention, the fact that it can still be around50%in some hard settings (as noted by the authors) suggests thatcomplete unlearningremains elusive. Future research needs to explore even more aggressive or fundamental ways to disentangle conflicting knowledge. - Generalization of : The
interpolation factoris ahyperparameter. Although the sensitivity analysis shows a robust window, its optimal value might vary across differentLLMsorrelation types. Automating this selection or making it less sensitive could be beneficial. - Computational Cost of Iteration: While
SPaEdit'sself-pacednature makes itscost allocation efficient, the increasing runtime per iteration (Figure 11) could become a bottleneck for extremely large models or massive sequential edits. Further work could explore more computationally efficient iterative solvers or approximations. - Complexity of Relation Editing: The paper categorizes
relation editingintonew relationandconditional relation. More complexrelation changes, such as those involvingmultiple subjectsorobjectsortemporal shiftsthat necessitate a re-evaluation ofmultiple related facts, might pose new challenges not fully covered here.
Transferability and Application:
The FE framework's unlearning mechanism could be highly transferable to any knowledge editing task where explicit forgetting of old facts is crucial, not just for relations. The self-paced learning approach is also a generalizable technique for improving the robustness of knowledge editing or even other fine-tuning tasks that suffer from difficult samples. This paper sets a new standard for how knowledge editing should be evaluated and performed, pushing the field towards more intelligent and responsible LLM maintenance.
Similar papers
Recommended via semantic vector search.