Papers
Topics
Authors
Recent
Search
2000 character limit reached

Dynamic Tokenization

Updated 28 January 2026
  • Dynamic tokenization is an adaptive framework that adjusts token boundaries based on input density and context, ensuring efficient and semantically meaningful segmentation.
  • It utilizes methods such as boundary prediction, adaptive clustering, and hierarchical merging to optimize token granularity across various modalities including language, vision, and genomics.
  • These techniques reduce computational load while preserving critical information, enabling robust performance and enhanced generalization in diverse modeling tasks.

Dynamic tokenization encompasses algorithmic frameworks and model architectures in which token boundaries and token sequence lengths are adapted in response to local input structure, content complexity, or task-specific signals, rather than being fixed a priori. The dynamic tokenization paradigm has been formulated across modalities—language, vision, multimodal (vision-language), genomics, and graphs—with the common goal of reducing redundancies, optimizing inductive bias, and facilitating efficiency and generalization by constructing tokens whose granularity is context-dependent.

1. Motivation and Conceptual Foundations

The primary motivation for dynamic tokenization stems from the inefficiencies and inductive limitations of static tokenization. Traditional tokenizers, such as Byte Pair Encoding (BPE) or fixed-grid visual patch extraction, create rigid and universal partitionings of input data, thereby over-fragmenting in high-density regions and under-segmenting in informative or sparse regions. This rigidity leads to sequence-length explosion in dense contexts (e.g., long videos), loss of semantic integrity (e.g., objects split across patches in images), over-fragmentation in low-resource or morphologically rich languages, or wasted modeling capacity on repetitive or uninformative genomic regions.

Dynamic tokenization frameworks adapt token boundaries and number based on intrinsic data properties—local density, semantic content, informativeness, motion, temporal regularity, etc.—and are instantiated via learned boundary prediction, adaptive clustering, hierarchical merging, or content-aware fusion (Zhang et al., 21 Mar 2025, Feher et al., 2024, Yan et al., 2024, Kim et al., 6 Jan 2026, Li et al., 17 Nov 2025, Wu et al., 2024).

2. Algorithmic Taxonomy: Techniques and Mechanisms

Dynamic tokenization methods can be grouped according to their operational mechanism and application domain. Core classes include:

  • Boundary Prediction and Learnable Segmentation: Token boundaries are predicted by a neural module (often an MLP, sometimes using Gumbel-Softmax relaxation) operating on learned representations (bytes, DNA bases, acoustic features), producing hard segmentations or soft probability masks. Tokens are formed as segments between predicted boundaries. Examples: FLEXITOKENS (Owodunni et al., 17 Jul 2025) (byte-level boundary prediction), DNACHUNKER (Kim et al., 6 Jan 2026) (H-Net dynamic DNA chunking), MergeDNA (Li et al., 17 Nov 2025) (iterative windowed token-merging).
  • Adaptive Clustering of Latent Embeddings: Tokenization operates by clustering local feature vectors, e.g., ViT patch embeddings, such that each cluster becomes a token. The cluster count adapts per input according to signal complexity. Approaches employ density-peak criteria, k-means, or other dynamic clustering rules. Notable instantiations: SeTok (dynamic density-peak clustering) (Wu et al., 2024), Token Dynamics (adaptive k-means over visual tokens) (Zhang et al., 21 Mar 2025).
  • Hierarchical Merging and Residual Quantization: Tokens are constructed by hierarchically merging adjacent or similar input fragments, often with multi-stage residual quantization (as in I²-World for 3D/4D scenes (Liao et al., 12 Jul 2025)). This is common in video and 3D world modeling, where both spatial and temporal redundancy can be exploited.
  • Content- and Motion-Aware Pruning: In video modeling, tokens are adaptively dropped based on motion estimation or redundancy (e.g., Gated Residual Tokenization (Zhang et al., 17 Sep 2025)), with static regions skipped and dynamic ones preserved.
  • Dynamic Graph Patchifying and Structure-Aware Tokenization: In graph domains, temporal or structural patches are created on-the-fly (Todyformer (Biparva et al., 2024)), with tokenization adapted to evolving subgraphs for balanced local/global modeling.

3. Mathematical Formulation and Architectural Patterns

The mathematical underpinnings of dynamic tokenization typically center on one or several of the following:

  • Boundary Prediction: Given an embedding sequence hth_t, boundary probabilities ptp_t are modeled as pt=σ(MLP(ht))p_t = \sigma(\mathrm{MLP}(h_t)) with boundary decisions sampled via hard-Gumbel-sigmoid (Owodunni et al., 17 Jul 2025, Kim et al., 6 Jan 2026).
  • Adaptive Clustering: Inputs tg,ft_{g,f} are clustered into KK centroids hkh_k, cluster assignment cg,f=argminktg,fhk2c_{g,f} = \arg\min_k \|t_{g,f} – h_k\|^2 with objective g,fminktg,fhk2\sum_{g,f} \min_k \|t_{g,f} – h_k\|^2 (Zhang et al., 21 Mar 2025, Wu et al., 2024).
  • Hierarchical (Residual) Merging: Multi-stage models apply successive local merging steps, performing operations of the form Zout=αZk+(1α)ZmZ_{\mathrm{out}} = \alpha Z_k + (1-\alpha) Z_m for selected pairs and updating a source matrix to track original assignment (Li et al., 17 Nov 2025, Liao et al., 12 Jul 2025).
  • Dynamic Masking: During training, suffixes of token sequences are randomly masked out (ElasticTok (Yan et al., 2024)), enforcing the model to reconstruct with variable-length codes and supporting dynamic token budgeting at inference.
  • Complexity Regularization: Compression rates and regularization are often controlled via loss terms—such as one-sided boundary penalties (Owodunni et al., 17 Jul 2025), ratio-based hinge losses (Kim et al., 6 Jan 2026), or batch-shaping for gating networks (Havtorn et al., 2023)—to ensure neither degenerate over-fragmentation nor collapse to trivially long tokens.

4. Empirical Performance and Efficiency

Dynamic tokenization methods consistently report substantial efficiencies with minimal trade-offs:

Method Relative Token Reduction Accuracy/F1 Loss Notable Benchmarks
Token Dynamics (Zhang et al., 21 Mar 2025) to 0.07% of original 1.13% absolute drop NextQA-MC
FLEXITOKENS (Owodunni et al., 17 Jul 2025) 15–30% seq. reduction up to +10 points F1/Acc XNLI, SIB-200, WikiANN
Dynamic Tokenization for LMs (Feher et al., 2024) >20% len. reduction <2 pp drop XNLI, UNER, MMLU
SeTok (Wu et al., 2024) tokens: 256→~19 +4–5% VQA, segmentation VQA, GQA, RefCOCOG
GRT (Zhang et al., 17 Sep 2025) 86% reduction in scenes Outperforms larger VLLMs DIVE (Dense Video QA)
MSViT (Havtorn et al., 2023) 20–40% tokens, ≈30–40% 0.3–0.7 pp gain/drop ImageNet, ADE20K
MergeDNA (Li et al., 17 Nov 2025); DNACHUNKER (Kim et al., 6 Jan 2026) Context-adaptive SOTA MCC/accuracy Genomics, NucleotideTransformer

Key findings demonstrate that adaptive tokenization:

5. Applications Across Modalities

Language Modeling

Dynamic tokenization methods such as FLEXITOKENS and on-the-fly retrofitting (Feher et al., 2024) enable adaptation to OOD languages, reduce subword fragmentation, and allow LMs to adapt their granularity post hoc. Approaches as in DNACHUNKER (Kim et al., 6 Jan 2026) and MergeDNA (Li et al., 17 Nov 2025) demonstrate end-to-end learnable segmentation of genomic sequences, robust to indels/shifts, and context-sensitive, outperforming previous k-mer or static subword baselines.

Vision and Video

Dense, high-resolution video understanding leverages dynamic tokenization to avoid quadratic explosion in token sequences. Gated Residual Tokenization (Zhang et al., 17 Sep 2025), ElasticTok (Yan et al., 2024), and Token Dynamics (Zhang et al., 21 Mar 2025) reduce token count based on local motion or redundancy, enabling feasible inference at high FPS with minimal loss in dense-sampling reasoning tasks. In image modeling, dynamic clustering (SeTok (Wu et al., 2024)) and mixed-scale gating (MSViT (Havtorn et al., 2023)) provide input-adaptive token counts, enhancing semantic segmentation and computational efficiency.

Multimodal and Graph Domains

Dynamic tokenization is instrumental in unifying representations across vision and language (LaVIT (Jin et al., 2023)), in temporal graph transformers (Todyformer (Biparva et al., 2024))—where local graph patches are dynamically segmented and tokenized, and in multimodal electronic health records, where temporal tokenization reflects the irregularities and multi-scale nature of real-world data (Ma et al., 2024).

3D and 4D Scene Forecasting

Advanced world models (I²-World (Liao et al., 12 Jul 2025)) combine hierarchical intra-scene dynamic tokenization via multi-scale residual quantization with inter-scene aggregation of temporal residuals, delivering tractable token streams for real-time autoregressive generation in high-dimensional, dynamic settings.

6. Theoretical and Computational Properties

Dynamic tokenization can be analyzed through:

  • Information-Theoretic Lenses: By aligning token granularity with local information density or entropy, these methods allocate representational bandwidth efficiently. Local density-peak clustering (SeTok) ensures that token count grows with semantic complexity, not input size (Wu et al., 2024).
  • Finite-State and Regular Transduction: Tokenization processes—both all-possible and canonical (e.g., MaxMatch, BPE)—can be framed as finite-state transductions, supporting efficient composition with regular constraints for guided generation (Cognetta et al., 2024).
  • Complexity Bounds: In most transformer architectures, the reduction of effective sequence length NN to Nˉ\bar{N} reduces compute from O(N2)O(N^2) to O(Nˉ2)O(\bar{N}^2). Adaptive token counts further multiply this efficiency when KNK \ll N as in (Zhang et al., 21 Mar 2025, Havtorn et al., 2023, Feher et al., 2024).
  • Semantic Integrity: Dynamic grouping around density peaks or semantic boundaries (SeTok, MSViT) can enforce tokens that better correspond to human-meaningful units, supporting interpretability and robustness.

7. Practical Considerations and Limitations

Advantages of dynamic tokenization include improved efficiency, adaptivity, semantic alignment, and in several cases, state-of-the-art empirical performance for a given compute budget (Zhang et al., 21 Mar 2025, Liao et al., 12 Jul 2025, Wu et al., 2024, Feher et al., 2024). Main limitations concern:

Dynamic tokenization represents a foundational shift in modeling strategies, replacing one-size-fits-all token boundaries with a unified, context- and content-adaptive approach. This enables efficient, semantically meaningful, and extensible modeling across the full spectrum of sequence learning applications.

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Dynamic Tokenization.