[Information Flow Routes: Automatically Interpreting Language Models at Scale 🔗](https://arxiv.org/abs/2403.00824)

Mapping the Mind of an LLM: How Information Flow Routes Reveal Model Inner Workings

The inner workings of Large Language Models (LLMs) often feel like a black box. We feed a prompt into one end, and a coherent response magically appears at the other. We know the architecture—Transformers, attention heads, feed-forward networks—but understanding exactly how a specific input token influences a specific output prediction remains one of the hardest challenges in AI research. Traditionally, researchers have tried to reverse-engineer these models using “circuits”—subgraphs of the model responsible for specific tasks. However, finding these circuits is usually a manual, labor-intensive process that requires human intuition to design specific test cases. ...

2024-03 · 11 min · 2142 words
[InfiniPot: Infinite Context Processing on Memory-Constrained LLMs 🔗](https://arxiv.org/abs/2410.01518)

InfiniPot: How to Fit Infinite Context into Finite Memory

The promise of Large Language Models (LLMs) often feels boundless, but in practice, it is strictly limited by memory. Whether you are summarizing a massive legal contract, analyzing a full-length novel, or maintaining a chat history that spans weeks, you eventually hit a wall: the context window. For cloud-based giants like GPT-4 or Claude 3, simply throwing more GPUs at the problem can extend this window to 100K or even 1M tokens. But what happens when we want to bring this intelligence to the “edge”—to our laptops and mobile phones? In these memory-constrained environments, we cannot simply add more RAM. When the input sequence grows too long, the application crashes or slows to a crawl. ...

2024-10 · 10 min · 1929 words
[Inference Helps PLMs' Conceptual Understanding: Improving the Abstract Inference Ability with Hierarchical Conceptual Entailment Graphs 🔗](https://aclanthology.org/2024.emnlp-main.1233.pdf)

Beyond Words: How HiCon-EG Teaches AI to Understand Concept Hierarchies

Introduction Imagine you read the sentence: “Mrs. Thompson gives her children some pasta.” As a human, your brain instantly performs a feat of abstraction. You understand that “pasta” is a type of “food.” Because you know she is giving “food,” you can infer a consequence: “The children are well-fed.” However, if you only viewed “pasta” as a physical object (an “entity”), the inference changes. If Mrs. Thompson gives an “entity,” the only safe inference is that the children “received an entity.” The nuance of feeding and being full is lost. ...

7 min · 1431 words
[InferAligner: Inference-Time Alignment for Harmlessness through Cross-Model Guidance 🔗](https://arxiv.org/abs/2401.11206)

Can We Make AI Safe Without Retraining? Meet InferAligner

The explosion of Large Language Models (LLMs) has shifted the landscape of artificial intelligence. We have moved past the era where only tech giants could run these models. Today, open-source base models like LLaMA and Vicuna are readily available, allowing developers to fine-tune them for specific domains—be it finance, medicine, or mathematics. However, this democratization comes with a significant catch: Safety. When you take a base model and fine-tune it on a specific dataset (say, medical records), you run the risk of “catastrophic forgetting” regarding its safety protocols. A model that was once polite and harmless might, after fine-tuning, be tricked into generating malware code or hate speech. Traditionally, fixing this requires training-time alignment—processes like Reinforcement Learning from Human Feedback (RLHF). But RLHF is expensive, complex, and computationally heavy. ...

2024-01 · 8 min · 1601 words
[Inductive-Deductive Strategy Reuse for Multi-Turn Instructional Dialogues 🔗](https://arxiv.org/abs/2404.11095)

How LLMs Can Learn to Ask Better Questions: The IDEAS Framework

Introduction In the rapidly evolving world of Large Language Models (LLMs), we often focus on how well a model answers a question. But there is another side to the coin that is equally critical for training these models: how well can a model ask questions? To align LLMs with human expectations, developers need massive datasets of high-quality, multi-turn dialogues. Manually collecting these conversations is expensive and slow. The solution? Use LLMs to generate the data themselves. One LLM plays the “System Agent” (the chatbot), and another plays the “User Simulator” (the human). ...

2024-04 · 10 min · 2048 words
[INDUCT-LEARN: Short Phrase Prompting with Instruction Induction 🔗](https://aclanthology.org/2024.emnlp-main.297.pdf)

Stop Writing Long Prompts: How INDUCT-LEARN Automates Prompt Engineering

If you have ever spent hours tweaking a prompt for a Large Language Model (LLM)—changing a word here, adding a constraint there, trying to get the model to “think” correctly—you have experienced the bottleneck of prompt engineering. We know that LLMs are capable of incredible reasoning, but their performance is often highly sensitive to the instructions they receive. While techniques like “Chain-of-Thought” (CoT) prompting significantly improve performance, they usually require humans to manually write out detailed reasoning steps. This is time-consuming and requires expertise. ...

7 min · 1363 words
[Incubating Text Classifiers Following User Instructions with Nothing but LLM 🔗](https://arxiv.org/abs/2404.10877)

Building Custom Text Classifiers from Scratch: How 'Incubator' Turns LLMs into Data Generators

Introduction Imagine you need to build a text classifier for a very specific task. Perhaps you need to filter emails that are both “urgent” and “related to shipping,” or identify social media posts that are “sarcastic” versus “genuinely angry.” Traditionally, you had two difficult options. First, you could collect thousands of examples and label them by hand to train a model—a slow, expensive process. Second, you could try “zero-shot” classification using raw text mining, which involves searching massive databases for keywords. However, this often fails when concepts are complex or nuanced. ...

2024-04 · 8 min · 1516 words
[Incomplete Utterance Rewriting with Editing Operation Guidance and Utterance Augmentation 🔗](https://arxiv.org/abs/2503.16043)

Teaching AI to Fill in the Blanks: A Graph-Based Approach to Incomplete Utterance Rewriting

Imagine you are texting a friend about a movie. Friend: “Have you seen Oppenheimer yet?” You: “Who is the director?” Friend: “Nolan.” You: “Oh, I love him.” To a human, this conversation is crystal clear. When you say “him,” you mean Christopher Nolan. When your friend says “Nolan,” they actually mean “Christopher Nolan is the director.” We constantly omit words (ellipsis) or use pronouns (coreference) because the context makes the meaning obvious. ...

2025-03 · 8 min · 1590 words
[In-context Contrastive Learning for Event Causality Identification 🔗](https://arxiv.org/abs/2405.10512)

How Contrastive Learning is Revolutionizing Event Causality Identification

How Contrastive Learning is Revolutionizing Event Causality Identification Causality is the bedrock of how humans understand the world. If we see a glass fall, we anticipate it might break. If we read that a heavy rainstorm occurred, we understand why the flight was delayed. For Artificial Intelligence, however, making these connections—specifically determining if one event explicitly caused another based on text—is a significant challenge. This task is known as Event Causality Identification (ECI). ...

2024-05 · 8 min · 1700 words
[In-Context Compositional Generalization for Large Vision-Language Models 🔗](https://aclanthology.org/2024.emnlp-main.996.pdf)

Beyond Simple Similarity: How to Teach Vision-Language Models to Generalize Compositionally

Introduction Imagine you are teaching a child what a “red apple” is. You show them a picture of a red apple. Now, you want them to understand a “green chair.” You show them a green chair. Finally, you present them with a “green apple”—an object they haven’t explicitly studied before, but which is composed of concepts they already know (“green” and “apple”). If the child recognizes it, they have demonstrated Compositional Generalization. ...

10 min · 2121 words
[In Search of the Long-Tail: Systematic Generation of Long-Tail Inferential Knowledge via Logical Rule Guided Search 🔗](https://arxiv.org/abs/2311.07237)

When LLMs Fail: Exploring the Long-Tail of Knowledge with Logic and Search

Large Language Models (LLMs) like GPT-4 and Llama 2 have dazzled the world with their ability to write code, compose poetry, and answer complex questions. But there is a catch: these models perform best when they are on “familiar ground.” When you ask an LLM about popular topics—like the iPhone or major historical events—it shines. But what happens when you push the model into the obscure corners of knowledge, known as the long-tail distribution? ...

2023-11 · 7 min · 1325 words
[Improving Spoken Language Modeling with Phoneme Classification: A Simple Fine-tuning Approach 🔗](https://arxiv.org/abs/2410.00025)

Can We Teach AI to Listen Like Humans? The Power of Phoneme Fine-Tuning

Introduction: The “Cocktail Party” Problem for AI Imagine you are at a loud, crowded party. Your friend is telling you a story. Despite the background music, the clinking of glasses, and the dozens of other conversations happening around you, you can perfectly understand what your friend is saying. You can strip away the noise, ignore the specific pitch of their voice, and focus entirely on the words and their meaning. ...

2024-10 · 9 min · 1886 words
[Improving Multi-party Dialogue Generation via Topic and Rhetorical Coherence 🔗](https://aclanthology.org/2024.emnlp-main.189.pdf)

Taming the Group Chat: How Reinforcement Learning Enhances Coherence in Multi-Party Dialogue AI

If you have ever been part of a busy group chat on WhatsApp or Slack, you know the chaos. Multiple conversations happen simultaneously. Someone answers a question from five minutes ago while two other people are debating lunch options. Keeping track of who is talking to whom—and more importantly, what they are talking about—is a significant cognitive task for humans. For Artificial Intelligence, this is a nightmare. In the field of Natural Language Processing (NLP), this problem is known as Multi-party Dialogue Generation (MDG). While standard chatbots (like early versions of Siri or simple customer service bots) only have to deal with one user (a one-on-one structure), MDG agents must navigate a web of entangled conversation threads. ...

11 min · 2193 words
[Improving Minimum Bayes Risk Decoding with Multi-Prompt 🔗](https://arxiv.org/abs/2407.15343)

Beyond the Perfect Prompt: How Multi-Prompt MBR Decoding Unlocks LLM Potential

Introduction If you have spent any time working with Large Language Models (LLMs), you have likely encountered the frustration of “prompt brittleness.” You spend hours crafting the perfect instruction, only to find that changing a single adjective or the order of examples drastically changes the output. This sensitivity is often seen as a bug, forcing engineers to hunt for a single “magic prompt” that solves a specific task. But what if we stopped trying to find the one perfect prompt? What if the sensitivity of LLMs to different instructions is actually a feature we can exploit? ...

2024-07 · 9 min · 1815 words
[Improving Knowledge Graph Completion with Structure-Aware Supervised Contrastive Learning 🔗](https://aclanthology.org/2024.emnlp-main.772.pdf)

Beyond the Triple: How StructKGC Teaches Language Models to See the Graph

Knowledge Graphs (KGs) are the silent engines powering much of the modern web. From Google’s Knowledge Vault to Wikidata, these massive networks store facts in the form of triples: (Head Entity, Relation, Tail Entity). For example, (Leonardo da Vinci, painted, Mona Lisa). However, KGs have a fundamental problem: they are never finished. Even the largest repositories suffer from incompleteness. This has given rise to the field of Knowledge Graph Completion (KGC)—the task of predicting missing links. ...

9 min · 1767 words
[Improving Discriminative Capability of Reward Models in RLHF Using Contrastive Learning 🔗](https://aclanthology.org/2024.emnlp-main.852.pdf)

Sharpening the Judge - How Contrastive Learning Fixes Reward Models in RLHF

Introduction In the current era of Generative AI, training a Large Language Model (LLM) to speak fluent English is effectively a solved problem. The frontier has shifted from capability to alignment. We don’t just want models that can write; we want models that write in accordance with human values—being helpful, harmless, and honest. The industry standard for achieving this is Reinforcement Learning from Human Feedback (RLHF). This technique fine-tunes models using a “Reward Model” that acts as a proxy for human judgment. Think of the Reward Model as a judge: if the judge has a keen eye and clear values, the AI learns to behave well. If the judge is confused, inconsistent, or easily tricked, the AI learns the wrong lessons. ...

10 min · 2013 words
[Improve Student's Reasoning Generalizability through Cascading Decomposed CoTs Distillation 🔗](https://arxiv.org/abs/2405.19842)

Breaking the Shortcut: How CasCoD Teaches Small Models to Reason Like Giants

In the rapidly evolving world of Artificial Intelligence, we are witnessing a “survival of the fittest” regarding model size. Large Language Models (LLMs) like GPT-4 possess an emergent ability known as Chain-of-Thought (CoT) reasoning. Instead of just jumping to an answer, they break down complex problems into intermediate steps, much like a human showing their work on a math test. However, running these massive models is expensive and computationally heavy. This has led to a surge in research focused on Knowledge Distillation—teaching smaller, more efficient “Student” models (SLMs) to mimic the reasoning capabilities of “Teacher” LLMs. ...

2024-05 · 8 min · 1687 words
[Improve Dense Passage Retrieval with Entailment Tuning 🔗](https://arxiv.org/abs/2410.15801)

Teaching Retrievers Logic: How Entailment Tuning is Solving the Relevance Gap in RAG

Teaching Retrievers Logic: How Entailment Tuning is Solving the Relevance Gap in RAG If you have ever built a Retrieval-Augmented Generation (RAG) system or an open-domain Question Answering (QA) bot, you have likely encountered a frustrating phenomenon: the “keyword trap.” You ask your system a specific question, like “Who was the first person to step on the moon?” The retriever goes into your vector database and pulls out a passage. But instead of an article about Neil Armstrong’s historic step, it retrieves a biography that says, “Neil Armstrong loved looking at the moon as a child.” ...

2024-10 · 9 min · 1825 words
[Impeding LLM-assisted Cheating in Introductory Programming Assignments via Adversarial Perturbation 🔗](https://arxiv.org/abs/2410.09318)

Can We Break ChatGPT? Preventing AI Cheating in CS Classrooms with Adversarial Attacks

Introduction The rapid rise of Large Language Models (LLMs) like ChatGPT and GitHub Copilot has fundamentally changed the landscape of software development. For professional developers, these tools are powerful productivity boosters. However, for computer science educators, they represent a looming crisis. In introductory programming courses (often called CS1 and CS2), the primary goal is to teach students the foundational logic of coding—loops, conditionals, and data structures. The problem? LLMs are exceptionally good at these standard problems. A student can copy a prompt, paste it into ChatGPT, and receive a working solution in seconds, bypassing the learning process entirely. ...

2024-10 · 8 min · 1692 words
[Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective 🔗](https://arxiv.org/abs/2407.02814)

Unmasking Bias in Vision-Language Models: Why Pictures Are the Real Culprits

Introduction In the rapidly evolving landscape of Artificial Intelligence, Vision-Language Models (VLMs) have become superstars. Models like CLIP or GLIP can look at an image and describe it, or read a text description and find the corresponding object in a picture. They are powerful tools, pre-trained on massive datasets of image-text pairs scraped from the internet. However, this power comes with a significant catch: societal bias. Because these models learn from human-generated data, they often inherit our stereotypes. For example, a model might be more likely to associate a “kitchen” with a woman or a “workshop” with a man, regardless of who is actually in the picture. ...

2024-07 · 9 min · 1727 words