Imagine teaching a smart assistant a new skill—say, recognizing your new puppy. It quickly learns to identify your dog, but in the process, it forgets who you are. This frustrating phenomenon, known as catastrophic forgetting, is one of the most persistent challenges in building AI systems that can learn continuously. The question is: how can we create models that adapt to new information without erasing what they already know?
This is the heart of Continual Learning (CL). Researchers have explored various strategies to tackle the problem, and two major families of methods have emerged:
- Regularization-based methods identify and protect the parameters most important for previous tasks, effectively “freezing” crucial knowledge.
- Meta-learning methods train the model not only to perform tasks but to learn how to learn, adjusting efficiently to new data while retaining old information.
Traditionally, these approaches have been treated as separate. However, the paper “Meta Continual Learning Revisited” bridges the gap and reveals that meta-learning is, in fact, implicitly performing a form of regularization through its use of higher-order information. Recognizing this connection allows the authors to pinpoint a core problem—high variance from noisy updates—and propose an elegant solution: Variance Reduced Meta-Continual Learning (VR-MCL).
In this article, we’ll unpack the ideas behind this work and explore how variance reduction can stabilize meta-learning, yielding a continual learner that is both adaptive and resilient.
Regularization and the Crucial Role of the Hessian
To understand the insight of this paper, we first need a solid picture of how regularization-based continual learning works.
When a model learns a new task, gradient descent updates its parameters to minimize the loss function. Unfortunately, these updates can interfere with parameters critical to prior tasks—thus the problem of forgetting.
Regularization-based CL methods add a penalty term to discourage large changes to “important” weights. But how do we decide which weights matter? The answer lies in the Hessian matrix, which captures the second-order derivatives of the loss and encodes the local curvature of the optimization landscape.
- High curvature directions (steep slopes) correspond to highly sensitive parameters—small changes in these can cause large loss variations. These must be preserved carefully.
- Low curvature directions (flat regions) allow for freer updates, enabling adaptation to new tasks.
In practice, the approximated loss across previous tasks is obtained through a Taylor expansion around their optimal parameters \( \hat{\theta}^i \). The resulting expression contains a Hessian term, dictating how strongly each parameter should be regularized.

Taylor approximation of the previous tasks’ loss function showing the Hessian component central to parameter regularization.
Minimizing this approximation leads to the following unified update rule used across regularization methods:

Unified parameter update rule for regularization-based continual learning algorithms.
In this formulation, we modify the gradient \( \nabla_{\theta} \mathcal{L}^j(\theta) \) by pre-multiplying it with the inverse of the accumulated Hessians from previously learned tasks. Each algorithm—EWC, IS, KFLA—differs mainly in how it approximates this cumulative Hessian.

Overview of continual learning methods under the unified Hessian approximation framework.
The limitation becomes clear when we note that these Hessians are fixed at the end of each task’s training. As learning continues and weights shift away from those original points, these static maps no longer accurately represent the landscape. It’s akin to navigating a city using an outdated map—each update increases the approximation error and weakens the model’s memory of prior knowledge.
Meta-Continual Learning: Implicitly Approximating the Hessian
Meta-Continual Learning (Meta-CL) approaches the problem differently. Instead of explicitly computing Hessians, it performs a bi-level optimization—an inner loop for task learning and an outer loop for meta-learning—supported by a memory buffer \( \mathcal{M} \) that stores examples from past tasks.
Here’s the process:
- Inner loop: The model quickly adapts to the current task using a few gradient steps.
- Outer loop: Using data from both the current task and the memory buffer, the model updates its parameters to perform well across all seen tasks. This update uses a hypergradient, effectively estimating second-order information.

The two-loop optimization structure of Meta-CL, showing the inner task adaptation and outer meta-update process.
Formally, we can write the optimization as:
\[ \min_{\theta} \mathcal{L}^{[1:j]}(\theta_{(K)}) \quad \text{subject to} \quad \theta_{(K)} = U_K(\theta; \mathcal{T}^j), \]where \( U_K \) represents \( K \) inner-loop gradient descent steps.
The authors’ theoretical analysis (Proposition 2) shows that the resulting update can be approximated as:

Iterative update rule showing that Meta-CL implicitly follows the regularization-based framework through online Hessian approximation.
This striking result reveals that Meta-CL performs the same type of weighted gradient update seen in regularization-based methods—but the Hessian is implicit, computed dynamically using data from the memory buffer. This adaptive nature allows Meta-CL to capture fresh curvature information, meaning its Hessian is always timely compared to the frozen estimates of regularization-based approaches.
However, this adaptability comes at a cost.
Since Meta-CL updates rely on random sampling from the memory buffer, the Hessian estimates can suffer from high variance. If the sampled data fails to represent certain tasks, the curvature for their important parameters may be underestimated, causing large destructive updates that lead to severe forgetting.
This leads to a crucial trade-off:
| Method Type | Hessian Accuracy | Variance | Adaptivity |
|---|---|---|---|
| Regularization | Low (static) | Low | None |
| Meta-CL | High (dynamic) | High | Strong |
The next step: can we achieve the adaptivity of Meta-CL while taming its variance?
Variance-Reduced Meta-Continual Learning (VR-MCL)
The proposed VR-MCL method does exactly this by integrating a momentum-based variance reduction technique into Meta-CL.
The key idea is intuitive: reduce the noise in gradient estimates by refining each update using information from the previous step. The rule for updating the variance-reduced hypergradient \( \hat{\mathbf{g}}_{\theta_b}^{\epsilon_b} \) at iteration \( b \) is:

Momentum-based variance reduction update formula used in VR-MCL.
Here’s what these terms mean:
- \( \mathbf{g}_{\theta_b}^{\epsilon_b} \): the current noisy hypergradient.
- \( \hat{\mathbf{g}}_{\theta_{b-1}}^{\epsilon_{b-1}} \): the corrected gradient from the previous iteration.
- \( \mathbf{g}_{\theta_{b-1}}^{\epsilon_b} \): gradient of the previous parameters on the current batch.
The correction term \( r(\hat{\mathbf{g}}_{\theta_{b-1}}^{\epsilon_{b-1}} - \mathbf{g}_{\theta_{b-1}}^{\epsilon_b}) \) acts as a control variate, smoothing out fluctuations caused by random sampling. The result is a much lower-variance update that stabilizes the optimization.

Flow of iterations in VR-MCL showing the use of historical information to reduce hypergradient variance.
Through theoretical derivations, the authors show that this variance reduction is mathematically equivalent to introducing a regularization term on the implicit Hessian. The stabilized update can be expressed as:

VR-MCL’s update rule under the new variance-reduced Hessian approximation.
This perspective clarifies how VR-MCL prevents destructive updates: it dampens movements along wrongly estimated low-curvature directions while maintaining cautious updates along high-curvature directions. By combining orthogonal strengths—timeliness from Meta-CL and stability from regularization—VR-MCL achieves both accuracy and resilience.
Experimental Evidence: Why VR-MCL Works
The effectiveness of VR-MCL is rigorously validated through extensive experiments across three standard online continual learning benchmarks: Seq-CIFAR10, Seq-CIFAR100, and Seq-TinyImageNet.
Overall Results

Average accuracy and anytime accuracy (AAA) on three continual learning benchmarks. VR-MCL consistently leads across all datasets.
VR-MCL significantly outperforms both regularization-based (On-EWC, IS) and meta-learning counterparts (MER, La-MAML), particularly on longer task sequences where forgetting is more severe. The performance gains demonstrate that variance reduction directly translates into better stability and continual retention.
Robustness to Buffer Size

VR-MCL maintains strong performance even at smaller buffer sizes, showing efficiency under restricted memory conditions.
In real-world streaming environments, memory capacity is limited. As shown above, VR-MCL consistently dominates across buffer sizes from 200 to 1000—an essential property for scalable continual learning.
Handling Imbalanced Data

VR-MCL remains robust against imbalanced data streams, outperforming specialized imbalance handling methods.
In imbalanced settings where task sample sizes vary drastically, most algorithms degrade sharply. VR-MCL’s variance reduction mechanism inherently mitigates this instability, maintaining high accuracy even under severe imbalance.
Measuring the Variance Directly

Relative gradient variance over training iterations: VR-MCL consistently shows reduced variance compared to standard Meta-CL.
This plot directly confirms the paper’s central hypothesis. As training progresses, VR-MCL maintains a lower gradient variance, leading to smoother optimization and more consistent knowledge retention.
Conclusion
The insights from “Meta Continual Learning Revisited” mark a meaningful advancement in continual learning research:
- Unified Understanding: Regularization-based and meta-learning approaches share a common foundation—they both rely on Hessian-based curvature information to balance learning new tasks with remembering old ones.
- The Hidden Challenge: Meta-CL’s strength in timely updates is offset by the high variance introduced through random memory sampling.
- The Solution: By integrating a momentum-based variance reduction mechanism, VR-MCL achieves a stable yet adaptable Hessian approximation, offering the best of both paradigms.
This work not only delivers empirical success but also clarifies the theoretical connection between two key schools of thought in continual learning. The findings suggest that mastering variance management is essential for enabling AI systems that can truly learn over a lifetime—continually, stably, and intelligently.
