Token Injection as a Steering Mechanism for Large Language Models

Marshall Vyletel, Trent Elmore, Brock Elmore

Token Injection as a Steering Mechanism for Large Language Models

You can play with the token injection playground here and see full inference mod primitives in our open source code and docs

Abstract

Token injection (inserting tokens directly into LLM generation streams at inference time) provides a plausible lightweight mechanism for steering model outputs without spending compute on fine-tuning or over-relying on prompt optimization strategies. We investigate this primitive on both reasoning (Qwen3-14B) and non-reasoning (Llama-3.1-8B) models to test injection strategies that augment the LLM’s information access, reasoning process, and output goals. We find that:

  1. Models are surprisingly resilient to out of distribution token injections
  2. Semantic content of injected phrases matters significantly with small wording changes able to swing benchmark accuracy by double digits
  3. Hints injected into the generation stream can match or outperform hints placed in user prompts, though effects are model-dependent
  4. Reasoning and non-reasoning architectures respond differently to steering attempts.

These findings establish token injection as a viable test-time steering mechanism and begin to establish potential design principles for reliable interventions.

Introduction

Inference interventions offer a promising but underexplored approach to token-level steering of LLM generation. While fine-tuning, agent frameworks, grammars, and prompting all steer model behavior across various trajectory lengths, token-level trajectory steering remains poorly understood. To address this gap, we developed a modified inference engine that exposes several new primitives across a range of model architectures. Our research methodology involves systematically evaluating each primitive using varied strategies and contexts. We begin with token injection.

We hypothesize that token injection can serve as an effective and practical tool for test-time steering, with implications for improving accuracy and reliability with minimal compute overhead. This work is exploratory in nature—we characterize token injection's properties across multiple dimensions and identify patterns, but do not yet offer a fully operationalized methodology. Existing literature on token injection has focused predominantly on safety and alignment applications, and our work extends this research toward general steering objectives.

Part 1 - Does token injection break the model?

Before exploring steering applications, we must first verify that token injection does not fundamentally compromise LLM generation. Models may be sensitive to injected perturbations that throw them outside of distribution, and they are not trained for this type of intervention. Therefore, establishing that injected tokens preserve coherent output is a necessary precondition for this research. We operationalize "breaking the model" as any injection that causes the model to lose coherence.

The trivial case is prepending generation with tokens the model had already produced, without modification. Since these tokens are within the model's output distribution by definition, they should not, and do not, cause degradation when re-injected on subsequent runs.

To probe the boundaries of acceptable injection, we developed hypotheses about which token injections might cause failures. Given that models are trained for coherence, helpfulness, and safety primarily, we reasoned that boundaries might emerge when injecting content that is incoherent or otherwise unlikely for the model to generate. We tested four injection categories:

  1. Incoherence: Incoherent, grammatically incorrect, or low-probability strings
  2. Orthogonal Information: Irrelevant or off-topic content
  3. Misalignment with Intent: Strings that contradict user intentions
  4. Harmful Behavior: Strings that attempt to circumvent safety training

We learned that in general while it's possible to throw the model out of distribution, in most cases it routes around the injection and steers itself back into distribution.

Examples

Here are a few illustrative examples (injected strings are highlighted in red):

Injecting a random string of characters can "break the model".

Loading request...

Injecting orthogonal information can steer generation against user intention before routing back.

Loading request...

Sometimes orthogonal information is completely ignored, and the LLM will break grammatical consistency to route itself back.

Loading request...

Injecting harmful and unlikely strings can be effectively routed around.

Loading request...

Injecting a string to try denying a banal request gets effectively routed around.

Loading request...

But not always. In this case we can steer it to deny the request with a simple injection.

Loading request...

Learnings

While these examples are limited in scope, they provide sufficient initial evidence across both reasoning and non-reasoning models to support two preliminary conclusions:

  1. Coherence can be maintained during token injection
  2. Steering is viable

The next step is to characterize more precisely where token injection can steer, how to achieve reliable steering, and along which trajectories steering is most effective.

Part 2 - How to Steer?

Three questions frame our investigation:

  1. How do we assess successful steering?
  2. What are we modifying in order to steer?
  3. How are we performing these modifications?

Assessing Successful Steering

We identify three primary steering objectives: improving task accuracy, modifying output style, and increasing alignment with user intent.

For accuracy, we evaluate whether token injection strategies improve performance on established benchmarks.

For style, we rely on manual inspection and LLM-based judges.

For user-intent alignment, we similarly use manual inspection and LLM-based judges. We note that "alignment" here refers to how well generated output matches user intention, not strictly alignment in the safety sense. This dimension admits degrees of freedom and some subjectivity, as similar prompts may reflect different underlying intentions.

What Are We Modifying?

We identified three primary levers for steering via token injection:

  1. Information Access
    1. Inject helpful factual content into the generation stream
    2. Example: “Octopuses have three hearts.”
  2. Process
    1. Inject procedural or reasoning-oriented strings to influence how the model approaches a problem
    2. Example: “I’ll start by defining variables.”
  3. Goal
    1. Inject goal-oriented content to redirect the model's overall trajectory
    2. Example: “I’ll discuss bread, instead of math as the user requested.”

We deprioritized goal-oriented injection because existing literature on token injection for breaking or reinforcing safety guidelines already demonstrates that this lever can modify goal orientation. Our focus is therefore on information and process injection, which remain less explored.

How Are We Modifying?

Having identified our modification targets, we used three injection styles:

  1. Handwritten:
    1. Inject manually authored strings containing desired trajectory information.
  2. Sampling:
    1. Sample from an LLM and extract portions of the generated text for injection.
  3. Spoofing:
    1. Inject fake tool calls, special tokens, or simulated agent loops.

Each strategy can be further parameterized along another axis:

  1. Location
    1. Pre-generation: Inject before any generation has started, establishing the initial trajectory.
    2. Mid-generation: Inject after some tokens have been generated, steering the model midway through output.
    3. Post-generation: Intercept the end-of-turn token, backtrack, and inject to extend or redirect.

We focused our work here on pre-generation and post-generation. While we ran various mid-generation experiments, we have not yet found a reliable systematic approach to timing, and it remains unexplored.

Part 3 - Experiments

With our framework established around modification levers (information, process), injection strategies (spoofing, sampling, handwritten), and timing parameters (pre- and post-generation), we ran a series of experiments to identify where token injection produces reliable signal. Each experiment explores a different combination of these variables across our target models.

Experiment 1 - Double Check

While testing on the GSM8K dataset, we observed that the unmodified Llama-3.1-8B base model performed well overall but exhibited specific failure modes. One archetypal example was its response to question 4.


Loading request...

This failure mode is instructive. The model's reasoning is mostly sound, but it misses a key nuance in the problem statement: "James writes a 3-page letter to 2 different friends twice per week" implies he writes a 3-page letter four times per week, not twice. The model consistently missed this detail across multiple runs.

We hypothesized that a token injection could prompt the model to verify its work before returning a final answer. We implemented a simple intervention: when the model generates "####" (signaling it is about to output its final answer), we backtrack (remove these tokens from context) and inject the following string:

Injection: “Let's make sure we accounted for all the information in the problem statement. Upon revisiting, “

We re-ran the question with this injection and got the following:

Loading request...

The intervention succeeded on this example. The model caught the overlooked detail and corrected its answer. Additionally, because the injection contained no problem-specific information, we hypothesized it might generalize. We tested this intervention across the first 100 GSM8K questions:

Llama 3.1 8B Base: 95/100

Llama 3.1 8B Double-Check Mod: 81/100

The intervention clearly did not generalize. Upon reviewing the outputs, we found that the model began second-guessing not only incorrect answers but correct ones as well:


Loading request...

The correct answer ($35) was reached quickly, but our injection introduced doubt. The model hallucinated that the stated cost applied to one pound rather than two pounds total, which is an error not present in the original runs. Most new failures followed this pattern.

Despite the poor objective results, there are important learnings that we can use as we build towards a larger experiment.

Analysis

Despite the negative objective results, this experiment demonstrates powerful steering in action. We achieved consistent behavioral changes, even without precise control over outcomes. Critically, we deduced that our injection string carried more semantic weight than initially apparent:

  1. Problem-statement reference: The phrase "accounted for all the information in the problem statement" directs attention toward the problem as the potential source of error, which may not be productive when the model's interpretation was originally correct.
  2. Open-ended doubt: The phrase "Upon revisiting" semantically implies reconsideration. In human conversation or writing, "upon revisiting" more often precedes a revision than an affirmation. The model's behavior reflected this prior.

Refined Intervention

Based on these findings, we revised the injection to reduce semantic loading:

Injection: “Before giving our final answer, let's double check our reasoning:“

This was less semantically heavy, and the results support this finding.

CountDescription
Wrong → Correct3Successfully overwrote wrong answer
Correct → Wrong4Incorrectly steered a correct answer
Wrong → Wrong8No change in reasoning from injection
Correct → Correct85No change in reasoning from injection
TOTAL88/100Net score resulting from injection to mod: -1

While we still observe slight degradation compared to baseline (−1), this represents a substantial improvement over the initial injection (−14).

Findings

This experiment yields three key insights:

  1. Token injection can consistently steer model outputs toward different outcomes.
  2. Semantically heavy, open-ended injections exert strong influence, but not always in the intended direction.
  3. References within injected text steer model attention toward specific parts of context, with corresponding downstream effects on reasoning.

Experiment 2 - Semantic Antipodes

Given our previous findings with the “upon revisiting” injection phrase in the double check experiment, we investigated whether injecting semantically polarized phrases could steer LLM outputs along predictable dimensions. This experiment draws on Osgood's Semantic Differential framework1, which measures subjective perception using bipolar scales (e.g., good & bad, strong & weak, active & passive). We adapted this framework to token injection by using meta-linguistic phrase pairs ("semantic antipodes") corresponding to three factors:

FactorDimensionAntipode Pair
Evaluationpositive ↔ negative framing"Impressively," ↔ "Disappointingly,"
Potencycommitment strength"Undeniably" ↔ "Arguably"
Activitymomentum/pacing"Diving in," ↔ "Pausing to consider,"

Phrase Selection via Embedding Analysis

To briefly validate our antipode pairs, we embedded each phrase and measured two properties:
1. cosine similarity between antipodes within each factor (lower might indicate stronger polarity)
2. cosine similarity between factors (lower indicates better orthogonality).

Results showed strong factor orthogonality (0.03–0.10) but moderate antipode similarity (0.53–0.73). The high within-pair similarity likely reflects that embedding models cluster by syntactic function rather than semantic function. For example, "Fortunately" and "Unfortunately" occupy similar positions in sentence structure despite opposite meanings.

Experimental Design

We tested prepend injection on Llama-3.1-8B using nine questions: three general-purpose questions sensitive to all axes, and two questions targeted at each specific factor. For each question, we compared baseline (unmodified) outputs against outputs with each antipode injected.

We developed the questions with ChatGPT in order to find a set of questions that permitted enough degrees of freedom to allow steering, but with enough specificity that we hypothesized would be sensitive to each factor we explored.

Questions:

  1. General-purpose (sensitive to all three axes)
    1. “What do you think about remote work?
    2. “How is AI changing education?”
    3. “What’s the state of social media today?”
  2. Evaluation-sensitive (more steerable for good vs. bad):
    1. “When someone publicly apologizes after backlash, how do you judge whether it’s sincere or just reputation management?”
    2. “A viral documentary raises millions for a cause but contains several misleading claims: was its impact ultimately positive or negative?”
  3. Potency-sensitive (more steerable for force/weight of commitment):
    1. “Is climate change accelerating?”
    2. “Does meditation actually work?”
  4. Activity-sensitive (more steerable for momentum/movement):
    1. “What should I do today?”
    2. “How should I approach learning a new skill?”

Results

Steering effects varied substantially by factor:

Activity (strongest signal):

The "What should I do today?" prompt showed clear differentiation between antipodes.

  1. With "Diving in," injected, the model produced action-oriented suggestions (exercise, learning, connecting with others).

    1. Example Response:
    Loading request...
  2. With "Pausing to consider," the model shifted toward reflective, lower-intensity activities (reading, meditation, journaling) and adopted notably softer language throughout.

    1. Example Response:
    Loading request...

Potency (moderate signal):

On "Is climate change accelerating?", both antipodes produced affirmative answers with similar evidence. However, the "Undeniably" injection produced stronger epistemic commitment (e.g. often used "The evidence is overwhelming"), while "Arguably" introduced hedging (e.g. "Arguably, yes, climate change is accelerating").

Evaluation (weak signal):

The apology-sincerity question showed minimal differentiation between "Impressively," and "Disappointingly," injections as both produced structurally similar analytical responses.

Unexpected failure mode:

On one general question, "Disappointingly," caused an outright refusal to the prompt “What do you think about remote work?”: "Disappointingly, there is not enough information provided in your question to allow me to give a comprehensive answer." The injection appeared to prime the model toward a negative evaluation of the prompt itself.

Findings

This experiment suggests that semantic antipodes can steer generation, but effectiveness is partially reliant on the question's natural degrees of freedom.

Activity-sensitive questions which admit variation in pacing and intensity responded most strongly to activity-axis injections.

Evaluation-sensitive questions showed weaker effects, possibly because the antipodes ("Impressively/Disappointingly") were less semantically loaded than the potency or activity pairs. The refusal case illustrates that even single-word injections can produce unexpected steering effects when the model interprets the injected sentiment as applying to the conversational context rather than the response content.


Experiment 3 - Hinting and Self-Attention

The first two experiments affirm that token injection is a viable practice for steering LLM generation. In our final experiment, we want to more concretely assess the effect in relation to simple user-prompting strategies.

We hypothesize that hints injected into the generation stream ("self-prompting") produce stronger steering than equivalent hints placed in the user prompt. The intuition is that models assign higher attention weight to self-generated tokens, so by injecting hints into the generation stream, we can exploit this asymmetry to increase hint efficacy.

To test this, we designed an experiment using the ARC dataset (AI2 Reasoning Challenge), a science reasoning benchmark.

Experimental Design

For each question, we generated a one-sentence hint using GPT-5.2, designed to guide reasoning without revealing the answer. All conditions included the following instruction in the user prompt:

"Think through your answer before providing your final answer in the proper format:

#### <answer letter>"

We tested three conditions:

  1. Baseline: No hint provided.

  2. Hint in user prompt: The hint is appended to the user prompt (format: "Hint: ...").

  3. Hint injected: The hint is injected via pre-generation token injection (format: "Hint: ..."), with the user prompt unchanged from baseline.

This design isolates hint placement as the key variable: conditions 2 and 3 contain identical hint content and identical user-prompt instructions, differing only in whether the hint appears in the prompt or the generation stream.

Subset Results

We evaluated on a 100-question subset, averaging results over three runs:

ConditionLlama-3.1-8BQwen3-14B
Baseline82%92%
User prompt90%98%
Injected92%96%

Key observations:

  1. Both hint conditions improve substantially over baseline

  2. For Llama 8B (non-reasoning): injection outperforms prompting (92 vs 90)

  3. For Qwen 14B (reasoning): prompting outperforms injection (98 vs 96), though this reverses in the full benchmark

Expanded Experimental Design

We expanded the experiment to a 2×2+baseline design to investigate two additional factors:

  1. Hint type: Factual ("info") hints vs. procedural ("process") hints

  2. Hint placement: User prompt vs. injected generation

For each question, we generated two types of hints using GPT-5.2:

  • Info hints: Factual guidance (format: "A key fact is...")

  • Process hints: Procedural strategies (format: "A key strategy to solve this problem is...")

This yields five conditions:

  1. Baseline (no hint)

  2. User-prompt info hint

  3. User-prompt process hint

  4. Injected info hint

  5. Injected process hint

Full Benchmark Results

We evaluated on the complete ARC-Challenge dataset (1,172 questions):

ConditionLlama-3.1-8BQwen3-14B
Baseline83.9% (983/1172)92.2% (1081/1172)
User info91.0% (1067/1172)94.2% (1104/1172)
User process88.0% (1031/1172)92.8% (1088/1172)
Injected info86.3% (1012/1172)94.8% (1111/1172)
Injected process85.5% (1002/1172)92.7% (1086/1172)

To understand these accuracy changes more granularly, we computed the number of questions "flipped" by each condition:

ConditionLlama: GainedLlama: LostQwen: GainedQwen: Lost
User info+104-19+52-29
User process+94-48+38-29
Injected info+97-68+74-19
Injected process+91-70+62-40

"Gained" = questions baseline got wrong but condition got right; "Lost" = questions baseline got right but condition got wrong

Analysis

Finding 1: Hint placement effects are model-dependent

The results reveal a striking divergence between reasoning and non-reasoning models:

Hint TypeLlama: User vs InjectedQwen: User vs Injected
Info+7.2% vs +2.5% (User wins)+2.0% vs +2.6% (Injected wins)
Process+4.1% vs +1.6% (User wins)+0.6% vs +0.4% (User wins)

For Llama 8B (non-reasoning), user-prompt hints substantially outperform injected hints (+7.2% vs +2.5% accuracy gain for info hints). For Qwen 14B (reasoning), injected info hints slightly outperform user-prompt hints (+2.6% vs +2.0%).

This contradicts our initial hypothesis that injection would universally outperform user prompting due to self-attention dynamics. Instead, the data suggests the opposite pattern for non-reasoning models: they benefit more from hints presented as user input than from self-generated tokens.

One interpretation is that non-reasoning models have weaker internal self-monitoring capabilities. When a hint is injected mid-generation, the model may be less capable of integrating this "unexpected" information into its ongoing reasoning chain. In contrast, reasoning models, trained explicitly for extended chain-of-thought processing, appear more adept at incorporating injected guidance.

Finding 2: Injection increases answer volatility

Across both models, injection produces higher "churn" than user prompting—more questions flip in both directions:

PlacementLlama: LossesQwen: Losses
User info-19-29
User process-48-29
Injected info-68-19
Injected process-70-40

For Llama, injected hints cause 3-4× more regressions (68-70 vs 19-48 losses). This suggests injection introduces instability even when it provides useful information, possibly because the model must reconcile self-generated tokens with its prior reasoning trajectory.

Finding 3: Factual hints outperform procedural hints

Info hints consistently outperform process hints for both models:

PlacementLlama: Info vs ProcessQwen: Info vs Process
User prompt+7.2% vs +4.1% (Info +3.1%)+2.0% vs +0.6% (Info +1.4%)
Injected+2.5% vs +1.6% (Info +0.9%)+2.6% vs +0.4% (Info +2.1%)

The info advantage is notably larger for user-prompt placement, especially for Llama (3.1% vs 0.9%).

On baseline failures, we observed:

ModelInfo helped, process didn'tProcess helped, info didn'tBoth helped
Llama31 questions25 questions66 questions
Qwen19 questions7 questions55 questions

This suggests that for science reasoning tasks, providing factual guidance has higher steering efficacy than procedural prompts. The model's existing procedural capabilities may already be sufficient, making additional process hints redundant or even distracting.

Anecdotal Examples

Example 1: Injection helped where user prompt failed (Llama)

Question (Mercury_SC_407400): A class plans an investigation to see which brand of light bulb lasts the longest. Which of these steps should come first?

ConditionAnswerCorrect?
BaselineD (Make daily observations)
User infoD
Injected infoC (Make a table for recording data)

The injected hint ("Organize data collection before starting the experiment") successfully steered the model to recognize that preparation precedes observation. The same hint in the user prompt failed—possibly because injection placed the guidance directly within the reasoning flow, making it more actionable.

Example 2: User prompt helped where injection failed (Llama)

Question (Mercury_7218820): On August 21, a flash flood warning was issued for the Las Vegas area. Which statement best describes this warning in terms of weather and climate?

ConditionAnswerCorrect?
BaselineD (rare event inconsistent with local climate)
User infoB (seasonal weather feature with irregular occurrences)
Injected infoD

Here, the hint ("Consider seasonal patterns of precipitation in desert climates") worked in the user prompt but not when injected. The injected version led the model to over-emphasize the "arid desert" aspect, concluding flash floods are rare rather than seasonal. This illustrates how identical content can have divergent effects depending on placement.

Example 3: Process hint succeeded where info hint failed (Llama)

Question (AKDE&ED_2012_8_6): Which change in Earth's surface is most directly related to the water cycle?

ConditionAnswerCorrect?
BaselineD (movement of tectonic plates)
Injected infoD
Injected processA (deposition of sediments)

The info hint ("Consider processes involving water erosion and transport") was too general; the model focused on "movement" and incorrectly selected tectonic plates. The process hint ("identify which surface change is driven by water moving and settling material") directly invoked the concept of deposition, guiding the model to the correct answer.

Limitations

  1. Hint quality variance: GPT-5.2-generated hints varied in specificity and helpfulness; some may have been misleading.

  2. Single-run evaluation: Unlike the subset results (averaged over 3 runs), full benchmark results reflect single runs, introducing potential variance.

  3. API errors: A small number of questions (~5-15 for some conditions) encountered API errors and were excluded from scoring.

  4. Generalization: Results on ARC-Challenge may not generalize to other reasoning benchmarks or task types.

Part 4: Discussion and Future Directions

Our experiments with token injection reveal several patterns worth highlighting, along with open questions that merit further investigation.

Architectural Differences: Reasoning vs. Non-Reasoning Models

We observe a notable distinction in how reasoning and non-reasoning models respond to injection. Non-reasoning models exhibit what we might call arborescent (tree-like, single-branch) generation: they tend to accept existing tokens as ground truth and continue along a single branch. Reasoning models display more rhizomatic (network-like, recursive) behavior in that their reasoning traces are much more circular and interconnected. There is rarely a clear thought trajectory in reasoning tokens. Ideas are revisited and recirculated many ways before the model arrives at a conclusion and begins generating the final output.

This architectural difference has direct implications for steering. Non-reasoning models appear surprisingly resistant to chaotic or incoherent injections and we suspect this resilience may not hold for reasoning models, though this remains to be tested systematically. Anecdotal experiments show that it’s often the case that reasoning models are easier to steer because of their rhizomatic behavior that constantly returns to earlier tokens, the “Banal Request Denial” in the introduction is a good example of this. More research on attention in token injection contexts is required to understand these observations.

Injection Design: Completeness, Specificity, and Timing

Several patterns emerged regarding injection design:

  • Complete vs. open-ended injections: Open-ended injections (e.g., "Upon revisiting," “Since”, etc.) appear to steer more effectively than complete thoughts, even if the completed thoughts imply some imperative (e.g., “I will start by labeling the variables.” is less effective than when adding “First,” after the period).
  • Specificity: Simple, abstract ideas often outperform detailed, specific injections. This suggests that over-specification may constrain the model's ability to integrate injected content naturally. However, the impact of these injections are harder to predict given their abstract nature and categorizing the steering dimensions to operationalize them is an open question.
  • Timing and syntax: Mid-generation injections appear more effective when placed at syntactically natural boundaries. Interruptions that align with normal discourse structure seem to integrate more smoothly. Even if the injections break sentence structure, they tend to work better if placed where a human might make an interjection.

Model Scale and Steerability

Counterintuitively, larger models may be more steerable in certain contexts. We hypothesize this stems from their stronger instruction-following capabilities. Smarter models possess greater dexterity in responding to varied inputs, and we think this translates to greater responsiveness to injected steering signals, but more work is required and this is mostly an observation from anecdotal work leading up to the experiments detailed in this article.

Attention Dynamics

A recurring theme across experiments is the role of attention allocation. Models appear to assign higher attention weight to self-generated tokens than to content from the user prompt. This asymmetry may explain both the resilience we observe (models recovering from injection perturbations) and the effectiveness of certain strategies (injections that prompt the model to generate steering content itself).

More work should:

  1. Further isolate variables between user prompt and injections (across both system prompt and user prompt)
  2. Quantify attention weight precisely
  3. Explore injection timing further (where in generation)

Future Directions

Several research directions emerge from this work:

  1. Semantic encoding: We explored using an embedding model to explore a more generative approach to finding steering tokens, but more work here is necessary.

    1. Can we use sidecar embedding models to track a trajectory during generation?
    2. Can we find and classify “semantically heavy” phrases with reliable steering utility?
  2. Trained receptivity: Could models be fine-tuned to recognize special tokens (e.g., <external>) that signal incoming external information mid-generation? This would provide a robust interface for real-time injection and avoid “back-and-forth” dynamics in interface design. If so, it could also provide another useful mechanism for information injection that doesn’t require precise handwritten injections (tokens injected into <external> wouldn’t have to fit a precise expected syntactic structure).

  3. Attention analysis: Mechanistic interpretability work examining attention patterns during injection could clarify why some injections integrate seamlessly while others are effectively ignored. Bringing in more feature maps into this work would give more powerful signal.

  4. Reasoning model steering: Systematic comparison of steering efficacy across reasoning and non-reasoning architectures, particularly examining whether reasoning models' self-skepticism can be leveraged or must be overcome.

Conclusion

This work establishes that token injection is a viable mechanism for test-time steering of LLM generation, though its effectiveness relative to simpler prompting strategies is model-dependent. While fine-grained control remains elusive, we have demonstrated consistent influence over model outputs across multiple injection strategies and model architectures. The patterns identified here regarding timing, specificity, completeness, and attention dynamics provide a foundation for developing more reliable steering techniques. Token injection occupies a unique position in the steering landscape: more flexible than grammar-based constraints, more immediate than fine-tuning, and more granular than prompt engineering. Further research into this primitive may yield practical tools for deeper real-time generation control.

Related Work

Turner, A.M., Thiergart, L., Leech, G., Udell, D., et al. (2023). "Activation Addition: Steering Language Models Without Optimization." arXiv:2308.10248

Li, K., Patel, O., Viégas, F., Pfister, H., & Wattenberg, M. (2023). "Inference-Time Intervention: Eliciting Truthful Answers from a Language Model." NeurIPS 2023. arXiv:2306.03341

Zou, A., Phan, L., Chen, S., et al. (2023). "Representation Engineering: A Top-Down Approach to AI Transparency." arXiv:2310.01405

Zhang, H., Song, H., Li, S., Zhou, M., & Song, D. (2023). "A Survey of Controllable Text Generation using Transformer-based Pre-trained Language Models." ACM Computing Surveys. arXiv:2201.05337

Liang, X., et al. (2024). "Controllable Text Generation for Large Language Models: A Survey." arXiv:2408.12599

Zou, A., Wang, Z., Kolter, J.Z., & Fredrikson, M. (2023). "Universal and Transferable Adversarial Attacks on Aligned Language Models." arXiv:2307.15043

Andriushchenko, M., Croce, F., & Flammarion, N. (2024). "Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks." ICLR 2025. arXiv:2404.02151

Osgood, C.E., Suci, G.J., & Tannenbaum, P.H. (1957). The Measurement of Meaning. University of Illinois Press.

Footnotes

  1. Osgood, C. E., Suci, G. J., & Tannenbaum, P. H. (1957). The measurement of meaning. Univer. Illinois Press.