<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://ijindal.github.io/feed.xml" rel="self" type="application/atom+xml" /><link href="https://ijindal.github.io/" rel="alternate" type="text/html" /><updated>2025-11-19T08:50:05-08:00</updated><id>https://ijindal.github.io/feed.xml</id><title type="html">Ishan Jindal</title><subtitle>personal description</subtitle><author><name>Ishan Jindal</name></author><entry><title type="html">Keeping LLMs Aligned Without the Cost: The Story of Instruction Residuals</title><link href="https://ijindal.github.io/posts/2025/10/instruction-residual-1/" rel="alternate" type="text/html" title="Keeping LLMs Aligned Without the Cost: The Story of Instruction Residuals" /><published>2025-10-10T00:00:00-07:00</published><updated>2025-10-10T00:00:00-07:00</updated><id>https://ijindal.github.io/posts/2025/10/instruction-reesidual</id><content type="html" xml:base="https://ijindal.github.io/posts/2025/10/instruction-residual-1/"><![CDATA[<p>Large Language Models (LLMs) constantly learn from new data. Yet, updating them often comes with a hidden cost: they forget how to follow instructions properly. Retraining to restore this ability is extremely costly.</p>

<p>What if we could skip that retraining altogether?</p>

<p>That’s the central idea behind my recent work, <em>“Keep the Alignment, Skip the Overhead: Lightweight Instruction Alignment for Continually Trained LLMs.”</em> This research introduces a simple yet powerful mechanism—<strong>Instruction Residuals</strong>—to retain and restore instruction-following behavior in LLMs without full instruction fine-tuning.</p>

<p>This work has been accepted at the <strong>ICML 2025 Workshop on Test-Time Adaptation: Putting Updates to the Test!</strong></p>

<hr />

<h2 id="the-core-idea-instruction-residuals">The Core Idea: Instruction Residuals</h2>

<p>Instead of re-aligning the entire model after every update, we extract the instruction-following <em>difference</em> between a base LLM and its instruction-tuned version. This <strong>instruction residual</strong> acts like a plug-and-play adapter.</p>

<p>Think of it as a “cheat sheet” of instruction-following skills that you can add back to your model after it learns new knowledge.</p>

<ul>
  <li><strong>Plug-and-play alignment:</strong> After updating your base model with new knowledge, simply add the residual to recover instruction capabilities.</li>
  <li><strong>Compute-efficient:</strong> Avoid repeating the entire instruction tuning process.</li>
  <li><strong>Modular and reusable:</strong> Works across architectures and model versions.</li>
</ul>

<hr />

<h2 id="why-this-matters">Why This Matters</h2>

<ul>
  <li><strong>Continual Pretraining is risky:</strong> Updating even a well-aligned LLM can degrade instruction-following ability by up to 10 points.</li>
  <li><strong>Instruction Residuals fix this:</strong> Restore—and often improve—instruction behavior with zero extra tuning.</li>
  <li><strong>Architecture-agnostic:</strong> Works across model families like LLaMa and Qwen.</li>
  <li><strong>Industry adoption:</strong> Already in use by:
    <ul>
      <li>DeepMind’s GAIA project within the GemmaVerse ecosystem</li>
      <li>CEIA-UFG’s Gemma-3-Gaia-PT-BR-4b-it multilingual model on HuggingFace</li>
    </ul>
  </li>
</ul>

<hr />

<h2 id="highlights-from-our-findings">Highlights from Our Findings</h2>

<ul>
  <li><strong>Restores instruction-following:</strong> After 1B tokens of continual pretraining, instruction residuals restore performance nearly to original levels, sometimes even exceeding them.</li>
  <li><strong>Cross-model portability:</strong> Residuals from LLaMa 3.1 can improve instruction behavior in LLaMa 3—showing backward compatibility.</li>
  <li><strong>Generalizes to domain-tuned models:</strong> Applied to domain-specific models like LLaMa-DocChat, instruction residuals boost instruction benchmarks by up to 6 points without harming domain QA ability.</li>
  <li>
    <h2 id="2000-less-compute-competitive-performance-a-detailed-analysis-revealed-that-traditional-instruction-tuning-demands-nearly-2000-more-flops-than-our-residual-approach-with-instruction-residuals-we-deliver-competitive-accuracy-on-benchmarks-like-mmlu-ifeval-and-gsm8k-at-a-fraction-of-the-computational-cost"><strong>2000× Less Compute, Competitive Performance:</strong> A detailed analysis revealed that traditional instruction tuning demands nearly <strong>2000× more FLOPs</strong> than our residual approach. With instruction residuals, we deliver competitive accuracy on benchmarks like <code class="language-plaintext highlighter-rouge">MMLU</code>, <code class="language-plaintext highlighter-rouge">IFEval</code>, and <code class="language-plaintext highlighter-rouge">GSM8K</code> at a fraction of the computational cost.</h2>
  </li>
</ul>

<h2 id="whats-next">What’s Next?</h2>

<p>This is just the beginning. Ongoing work focuses on:</p>

<ul>
  <li>Extending instruction residuals to smaller models (≤1.5B) without quality loss.</li>
  <li>Generalizing across model architectures like Mistral, Mixtral, and future Qwen variants.</li>
  <li>Exploring residual approximation techniques when both the base and instruction-tuned models are not available.</li>
</ul>

<p>Reducing instruction tuning costs could democratize LLM alignment, making advanced models more accessible to researchers and smaller organizations.</p>

<hr />

<h2 id="-acknowledgments">🙏 Acknowledgments</h2>

<p>Thanks to my collaborators at Samsung R&amp;D Institute India, and the broader research community for supporting this work. The momentum we’ve gained—both in academia and industry—has been incredible.</p>]]></content><author><name>Ishan Jindal</name></author><category term="LLMs" /><category term="Instruction-residual" /><category term="Continuous pre-training" /><summary type="html"><![CDATA[Large Language Models (LLMs) constantly learn from new data. Yet, updating them often comes with a hidden cost: they forget how to follow instructions properly. Retraining to restore this ability is extremely costly.]]></summary></entry></feed>