Attribution of performance gains to fine-grained task decomposition

Determine whether the observed outperformance in the hierarchical multi-agent large language model trading system is fundamentally attributable to the fine-grained task decomposition of agent prompts, rather than to confounding factors such as large language model vocabulary-preference effects that may propagate to downstream agents.

Background

The paper proposes a hierarchical multi-agent LLM trading framework and shows that specifying fine-grained, analyst-like tasks (especially for Technical and Quantitative agents) improves backtested Sharpe ratios compared to coarse-grained prompts. Ablation and text-similarity analyses suggest that technical signals propagate more effectively under fine-grained instructions.

However, the authors note uncertainty about the causal mechanism behind the performance gains. They highlight a potential alternative explanation: certain vocabulary patterns preferred by LLMs might unduly influence downstream agents, implying that linguistic biases rather than genuine task decomposition effects could drive the observed improvements.

References

First, it is not yet fully clear whether the performance gains are fundamentally attributable to fine-grained task decomposition itself. One alternative explanation is that certain vocabulary patterns may be more easily adopted by the preference of LLMs to influence downstream agents.

Toward Expert Investment Teams:A Multi-Agent LLM System with Fine-Grained Trading Tasks  (2602.23330 - Miyazaki et al., 26 Feb 2026) in Discussion and Conclusion — Limitations and Future Work