Papers
Topics
Authors
Recent
Search
2000 character limit reached

Character-Level LLMs

Updated 29 December 2025
  • Character-Level LLMs are neural architectures that operate on individual characters, enabling precise tasks such as spelling correction and intra-token manipulation.
  • Hybrid designs combine character-level and token-level processing to mitigate tokenization-induced bottlenecks using specialized modules and auxiliary objectives.
  • Empirical benchmarks show that targeted methods like reverse prediction, hierarchical RNNs, and character-enhanced Transformers significantly improve character-level task performance.

Character-level LLMs are neural architectures and training paradigms designed to explicitly represent, reason over, or generate text at the granularity of individual characters, rather than being restricted to word-level or subword-token-level processing. While word- and subword-tokenized LLMs dominate contemporary natural language processing, mounting evidence demonstrates that they struggle with basic tasks requiring character-level understanding, such as string manipulation, spelling correction, and intra-token position prediction. This has motivated the development of new architectures, training objectives, and evaluation frameworks centered on character-level reasoning, with the goal of overcoming systematic limitations imposed by traditional tokenization.

1. Tokenization-Induced Limits on Character-Level Reasoning

Subword tokenization, such as byte-pair encoding (BPE) or WordPiece, enables efficient text compression and shortens sequence lengths in LLMs, but it systematically obscures the internal character structure of words and phrases. Investigations demonstrate that standard LLMs can spell out token strings with high accuracy under dedicated prompting, but fail at more granular character-level queries—such as counting characters, extracting the kkth character, or conducting intra-token substitutions—because only the first character of a token is reliably encoded at the embedding layer. Downstream Transformer layers must reconstruct or “decode” character-level structure—an ability that emerges only in specific mid-to-upper layers, and only after substantial pretraining. Probes reveal a sharp “breakthrough” effect, after which character signals become explicit and manipulable, mediated by specific “knowledge neurons” and attention routing, but this signal is not uniformly present across the architecture (Hiraoka et al., 12 Jun 2025).

The mutual information framework formalizes the bottleneck: given a subword token WW and its character sequence CC, human readers know H(CW)=0H(C|W)=0; however, in large, naturally tokenized corpora, the context XX contains little information about CC, so the model receives almost no gradient for character-level facts. A percolation-theoretic analysis shows that character-level capabilities emerge suddenly and late during training, only after a critical mass of token–character pairings are encoded (Cosma et al., 20 May 2025).

A series of benchmarks, such as CharBench, further demonstrates that the length of the token containing the queried character is the strongest negative correlate of success on position-based character-level tasks, while token count or compression ratio shows weaker or no correlation (Uzan et al., 4 Aug 2025). The effect is universal: LLMs of all scales and designs—including GPT-4, Llama 3, and open-weight models—exhibit striking deficits in basic character-level accuracy compared to token-level tasks, despite human-level performance from even non-expert annotators (Shin et al., 2024).

2. Character-Level Modeling Architectures and Hybrid Designs

Several architectural strategies target the character-level blind spot of tokenized LLMs, ranging from pure character-level LLMs to multi-level hybrids:

Character-Word LSTM Models

The Character-Word LSTM model concatenates fixed-length character embeddings for each input word with standard word embeddings. By tuning the number of character slots and embedding size, the model preserves input dimensionality while improving perplexity over baseline word-level LMs and reducing out-of-vocabulary (OOV) errors. Ablations indicate the improvements stem from real character information rather than regularization or random noise. Parameter sharing among character embeddings further reduces footprint with only minor degradation. Rare or OOV word modeling is enhanced because character-level channels encode subword structure (Verwimp et al., 2017).

Hierarchical Character-Level RNNs

Hierarchical RNNs introduce multi-timescale processing: a fast character-level module and a slower word-level module. The character RNN operates at every step, and is reset or conditioned at word boundaries by information from the word-level RNN. This design achieves lower word perplexity than deep monolithic LSTM LLMs with a fraction of the parameters, and is especially effective for speech recognition and sequence modeling with strict OOV support (Hwang et al., 2016).

Character-Enhanced Transformers

Recent interventions for Transformer LLMs inject character-level encoders into the token processing pipeline. The “block-causal” cross-attention approach introduces a 1-layer Transformer that encodes all characters present in each token into grouped “blocks,” and each main layer of the token Transformer receives cross-attentive signals from these character representations. This enhances character-level task performance by over an order of magnitude and eliminates dependence of emergence timing on vocabulary size. Crucially, this design does not sacrifice token-level performance or efficiency and provides a direct architectural path for fine-grained reasoning (Cosma et al., 20 May 2025).

Output Head Decomposition for Generation

SpeLLM circumvents the quadratic cost of large token projection heads by decoupling the input and output vocabularies: the model generates multiple characters per step via kk independent head projections of size O(sd)O(s \cdot d), where ss is the size of the character set and dd the hidden state dimension, instead of a full BPE vocabulary projection. Trained by distillation from a standard LLM, SpeLLM matches or exceeds teacher accuracy on many downstream benchmarks with a 5% average runtime speedup, while supporting output for rare or underrepresented scripts (Ben-Artzy et al., 22 Jul 2025).

3. Supervision Paradigms for Character-Internal Structure

Token Internal Position Awareness (TIPA) introduces auxiliary training tasks, such as reverse character prediction, in which the model must output character-identity/index pairs in descending order given a token. This objective tightly couples internal position information with token-level representations, leading to markedly improved performance in tasks such as Chinese Spelling Correction, both where character position must be explicitly predicted and in standard correction settings. The benefit holds for both single-token and multi-token variants (MTIPA), and reverse ordering aids learning of token length. TIPA and MTIPA require no modification to the core architecture or tokenizer, and facilitate broader cross-lingual applicability for languages with rich subtoken structures (Xu et al., 2024).

For languages where character-level constraints are operationally essential, such as Chinese, dedicated character-level tokenization coupled with continued pretraining and targeted fine-tuning (as in C-LLM) restores one-to-one sequence alignment and enforces hard constraints like length preservation and phonetic similarity. C-LLM yields substantial improvements on Chinese spelling correction benchmarks and nearly eliminates length and non-homophonic mismatch errors, which persist in mixed tokenization settings (Li et al., 2024).

4. Divide-and-Conquer Methodologies for Inference-Time Manipulation

To circumvent LLMs’ inability to perform atomic character manipulations due to tokenization, a divide-and-conquer methodology decomposes operations (deletion, insertion, substitution) into three stages:

  1. Token Atomization: Render each input word as a sequence of individual character tokens with explicit separators to "unlock" otherwise implicit character reasoning states.
  2. Per-Character Manipulation: Perform the requested transformation positionally (e.g., deleting, inserting, or substituting characters), leveraging the LLM’s ability to “spell out” characters reliably when atomized.
  3. Controlled Token Reconstruction: Reconstitute the target sequence, one subword at a time, using the model's token probabilities (i.e., reconstruct the word by greedy BPE-hypothesis selection given the character sequence).

This approach leads to immediate, dramatic accuracy gains—up to 73.9% improvement for insertion tasks without further training—across a range of major LLMs and is robust to prompting style and shot count. However, residual error modes include autocorrect bias and early stopping in multi-target edits (Xiong et al., 12 Feb 2025).

5. Empirical Benchmarks and Analytical Frameworks

Quantitative evaluation uniformly shows substantial performance gaps between character-level and token-level tasks. In CharBench, with over 40,000 stratified character-level questions, state-of-the-art LLMs only reach average accuracies of 43.6% on first-occurrence indexing and 32.3% on last-occurrence, compared to more than 70% on token-level analogues. Model success is strongly and negatively correlated with the size of the target token for position-based reasoning, but only weakly affected by word-token counts or gross compression ratios. Counting accuracy declines most with the absolute occurrence number rather than tokenization properties (Uzan et al., 4 Aug 2025).

Probing studies further reveal positional encoding for only the first character at the embedding layer, with downstream layers—which correspond to spikes in both probe accuracy and attention focused on the target token—“reconstructing” deeper internal structure, supporting the notion of a nontrivial “breakthrough” stage for character-level logic (Hiraoka et al., 12 Jun 2025). Synthetic task suites, as in “The Strawberry Problem,” show a phase transition where character-level competence emerges suddenly and late, with patterns consistent with percolation theory and strongly tied to vocabulary size and token length (Cosma et al., 20 May 2025). These analyses suggest not only the necessity of architectural change for truly fine-grained character modeling, but also a predictive, quantifiable basis for understanding when and why failures arise.

6. Implications, Limitations, and Future Directions

Systematic failures in character-level tasks stem directly from the mutual-information bottleneck imposed by tokenization, rather than only model scale or architectural sophistication. While richer supervision schemes (e.g., TIPA), architectural hybrids with side-channel character blocks, and output-head factorization mitigate these deficits, none fully resolve them without explicit character-level structure in modeling and training. Benchmarks demonstrate that increasing vocabulary size exacerbates the problem and delays emergence of character-level competence (Cosma et al., 20 May 2025).

Future models may benefit from hybrid token–character transformer architectures, explicit multi-granular embedding strategies, or integration of visual/textual character features for scripts and domains where compositional spelling is critical (Shin et al., 2024). Auxiliary objectives that force position prediction or reverse spelling, adaptive tokenization schemes, and modular “character processors” for interpretability and targeted fine-tuning are active areas for research. Crafting training curricula that interleave character- and token-level losses may induce faster, earlier generalization, while specialized benchmarks like CharBench provide actionable, diagnostic signal for model development and automated tokenizer tuning (Uzan et al., 4 Aug 2025).

In summary, character-level LLMs address a fundamental limitation in standard tokenized LLMs, offering advances in character manipulation, spelling correction, and robustness to OOV or novel forms without sacrificing subword modeling efficiencies. Technical progress in this domain is grounded in a combination of architectural innovation, supervised curriculum design, and precise empirical evaluation, with broad implications for languages, tasks, and applications that require fine-grained text understanding.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Character-Level LLMs.