ProAct-Helper: AI-Driven Proactive Assistant
- ProAct-Helper is an AI-driven proactive agent that minimizes human task cost by synthesizing assistant policies based on user belief updates and formal cost/recognition constraints.
- It employs a two-level Monte Carlo Tree Search combined with conformant planning to optimize joint action strategies under uncertainty.
- Experimental results demonstrate that ProAct-Helper can reduce human solo costs by 35–70% and significantly lower cognitive workload in collaborative settings.
A ProAct-Helper is an AI-driven agent or system designed to provide proactive task assistance, typically in environments where humans and AI agents co-exist with only partial mutual observability and potentially limited shared intent. The core challenge is to synthesize agent behavior that not only reduces the human's cost-to-go for a target task but also makes this benefit both practically effective and cognitively accessible to the human, even when the human may not expect, request, or fully recognize the need for assistance. These helpers are formally motivated and evaluated in frameworks that integrate planning under uncertainty, user-belief modeling, and principled cost/recognition constraints. The term has gained currency in work on collaborative planning, mixed-initiative interaction, and the evaluation of proactive assistant architectures (Kulkarni et al., 2021).
1. Formal Environment, Objective, and Recognition Constraints
A canonical ProAct-Helper operates in a stochastic environment with a set of state variables , human and robot (assistant) action spaces and , and a transition function encoding the effects of each action. Each human action and assistant action is described by preconditions, add/del effects, and an immediate cost or . The human perceives a finite observation space , with mapping action-state pairs to observation symbols (including the null for unperceived events). The assistant is assumed to have full state observability.
The core optimization is to synthesize an assistant policy mapping beliefs to actions so as to minimize the overall human cost-to-go under three core principles (Kulkarni et al., 2021):
- Cost reduction: The assistant's actions must decrease the human's total cost towards their goal, i.e., for any joint plan ,
where is the optimal solo plan.
- Recognition: The human must be able to recognize this cost reduction, formalized as a recognition constraint for some ,
at the assistant's hand-off time and human's belief .
- Overall cost-optimality: The assistant must optimize the human's expected cost (including workload and time), as measured after the human belief update.
In this mixed-observability framework, the human's belief is updated by a Bayesian filter,
with identity update on null observations.
2. Proactive Planning Algorithms and Policy Computation
Policy synthesis for a ProAct-Helper employs a two-level Monte Carlo Tree Search (MCTS) architecture combined with conformant planning (Kulkarni et al., 2021). The high-level pseudocode structure incorporates:
- Outer-loop MCTS over assistant actions and beliefs.
- Belief updates based on simulated human perceptions.
- Leaf evaluation by conformant planning (e.g., Conformant-FF), which estimates the human's minimum residual cost from the current belief.
- Only positive rewards propagate for simulated joint plans satisfying .
The backpropagated reward at each node is formed from the combined cost surrogate
with a tunable weight. Only policies satisfying a user-specified recognition constraint are considered valid.
Following tree construction, a greedy rollout searches down the top-k utility children per level, selecting the assistant plan that minimizes this cost surrogate subject to the recognition constraint.
3. Experimental Evaluation and User Studies
Empirical evaluation covers both simulation-based benchmarks and user studies:
- On empirical domains (e.g., USAR, Driverlog), proactive assistance reduced human solo cost by 35–70%, with assistant plans remaining within a fixed step-budget (e.g., steps), and planning times under 20 s with MCTS iterations.
- User studies (n≈30) found that without explicit legible or obfuscating display actions only 20% of participants recognized the proactive assist. Introducing a single legible or obfuscating action increased recognition to 80–90%, validating the necessity of making cost reduction “legible."
- Workload and processing-load ratings (NASA-TLX, 1–7 scale) saw substantial drops when ProAct-Helper plans were used: workload increases of +2.7 (baseline to assistant), effect sizes .
The results support (a) sufficiency of a single explicit action for recognition and (b) large reductions in workload and cognitive load, validating the composite cost and recognition framework (Kulkarni et al., 2021).
4. Integration of ProAct-Helper with Broader Proactivity Architectures
The ProAct-Helper abstraction appears as a recurrent motif in subsequent proactive agent research. The PROPER framework (Kaur et al., 14 Jan 2026) splits proactivity into two agents: a Dimension Generating Agent (DGA) that surfaces implicit knowledge gaps and a Response Generating Agent (RGA) that integrates both explicit and implicitly inferred needs in system responses. In both, proactive assistance is formalized by surfacing unarticulated but contextually salient dimensions and calibrating the initiative (avoiding both underreach and overreach), with explicit scoring on coverage, initiative appropriateness, and user-intent alignment.
These architectures operationalize ProAct-Helper’s principles outside classical planning settings—in LLM-based language or task assistants, proactive programming systems (Chen et al., 2024), and LLM-powered visual analytics (Zhao et al., 24 Jul 2025)—by maintaining explicit representations of observed and inferred user needs, and proactively suggesting actions or modifications that can be directly integrated by the user.
5. Generalizations, Limitations, and Future Research Directions
Current ProAct-Helper frameworks rely on formal task models (fluents, transition functions) and explicit modeling of user beliefs and observability. Limitations acknowledged in (Kulkarni et al., 2021) include:
- The models' reliance on full state enumerability and tractable belief update may limit scalability in high-dimensional, continuous, or partially known domains.
- Recognition constraints presuppose that a single explicit action suffices to induce belief update for the human agent, which can break down if users ignore, misinterpret, or forget such cues.
- Cost models are constructed for single users; generalization to multi-human teams remains an open line.
Research directions include richer user models (memory, persistence, social context), dynamic adaptation of recognition thresholds based on observed human feedback, and integration with large-scale, partially observable LLM-based environments as in (Kaur et al., 14 Jan 2026) and (Lu et al., 2024). Cross-domain extensions (robotics, design, programming, safety) increasingly couple ProAct-Helper mechanisms with multimodal and hybrid architectures (DNN+symbolic, task graph planners, reward models) to maintain proactive, interpretable, and intent-aligned assistive behaviors.
6. Summary Table: ProAct-Helper—Core Model Components
| Component | Formalization | Purpose |
|---|---|---|
| State Space | instantiations of fluents | Describes all possible world states |
| Actions (H, R) | , , with , , | Available moves and their costs |
| Transition Function | , | Dynamics of environment |
| Observation Model | , | What the human perceives per action |
| Belief Updates | Captures human's world belief | |
| Assistant Policy | Maps beliefs to actions | |
| Objective | Minimize human cost, ensure recognition, optimize total step cost | Satisfies proactive principles (Kulkarni et al., 2021) |
The ProAct-Helper concept thus formalizes and operationalizes proactive, legible, and cost-optimal AI assistance under uncertainty, combining structured planning, user model updating, and empirical validation in diverse collaborative settings (Kulkarni et al., 2021).