Papers
Topics
Authors
Recent
Search
2000 character limit reached

Agentic Business Process Management: A Research Manifesto

Published 19 Mar 2026 in cs.AI | (2603.18916v2)

Abstract: This paper presents a manifesto that articulates the conceptual foundations of Agentic Business Process Management (APM), an extension of Business Process Management (BPM) for governing autonomous agents executing processes in organizations. From a management perspective, APM represents a paradigm shift from the traditional process view of the business process, driven by the realization of process awareness and an agent-oriented abstraction, where software and human agents act as primary functional entities that perceive, reason, and act within explicit process frames. This perspective marks a shift from traditional, automation-oriented BPM toward systems in which autonomy is constrained, aligned, and made operational through process awareness. We introduce the core abstractions and architectural elements required to realize APM systems and elaborate on four key capabilities that such APM agents must support: framed autonomy, explainability, conversational actionability, and self-modification. These capabilities jointly ensure that agents' goals are aligned with organizational goals and that agents behave in a framed yet proactive manner in pursuing those goals. We discuss the extent to which the capabilities can be realized and identify research challenges whose resolution requires further advances in BPM, AI, and multi-agent systems. The manifesto thus serves as a roadmap for bridging these communities and for guiding the development of APM systems in practice.

Summary

  • The paper presents Agentic BPM, extending traditional BPM by integrating framed autonomy to align autonomous actions with organizational objectives.
  • It details a modular architecture that separates macro-level process framing from micro-level agent execution, enabling effective multi-agent collaboration.
  • It outlines research challenges in explainability, conversational actionability, and self-modification, establishing a roadmap for interdisciplinary advancements.

Agentic Business Process Management: Foundations, Capabilities, and Research Challenges

Introduction and Motivation

The proliferation of AI technologies—including advanced analytics, machine learning, and LLMs—has created new opportunities and risks for organizational software systems, particularly in the context of Business Process Management (BPM). While automation has traditionally driven BPM, a shift toward autonomous, agent-based architectures poses significant governance and alignment challenges: autonomous software agents risk making non-compliant decisions, creating functional uncertainties, and introducing new technical debt if their behaviors are unregulated.

This manifesto introduces Agentic Business Process Management (APM) as a conceptual extension of BPM to address these challenges. In APM, autonomous agents—both human and artificial—become first-class functional entities. Central to APM is the notion of process awareness, realized via "framing," which constrains and aligns agent autonomy with organizational objectives. The paper methodically articulates APM's conceptual underpinnings, delineates architectural blueprints, formalizes required agent capabilities, and offers a structured agenda of research challenges.

Conceptual Foundations and Architecture

Core Abstractions

APM is conceptualized as an agentic, socio-technical system. Agents in APM are autonomous, proactive functional units capable of perceiving, reasoning, and acting to achieve their own goals, yet their autonomy is framed to ensure alignment with organizational process constraints and objectives. Agents may be human, AI-based, software-only, or embodied (robotic).

Process awareness is foundational: it requires that agent behavior coheres with explicit organizational processes, constraints, regulations, and collective goals. Critically, APM distinguishes autonomy from automation; where automation follows rigid, predefined workflows, framed autonomy in APM enables context-sensitive agent decision-making constrained by process-awareness.

System and Agent Architecture

The conceptual architecture of an APM system structurally separates macro-level (system) framing from micro-level (agent) execution: Figure 1

Figure 2: Conceptual architecture, distinguishing between macro-level process framing and micro-level agentic perceive–reason–act cycles, with process-awareness enforced via framing mechanisms.

At the macro-level, collective process goals are established, and process awareness is enforced through explicit framing mechanisms, which impose normative (often deontic) restrictions over agent goals and knowledge states. At the micro-level, each process-aware agent independently executes perceive–reason–act loops over its framed internal representations, interacting with both human and artificial agents, tools, and external services.

This architectural blueprint is modular and extensible: agents are equipped with tools—sensors, actuators, messaging, external services—that augment their ability to interact with their environment and with each other. Shared ontologies or communication protocols support interoperable interactions, essential for coherent multi-agent collaboration.

Envisioned Core Agent Capabilities

The manifesto specifies four essential agent capabilities within APM:

Framed Autonomy

Framing introduces explicit, evolving sets of rules, constraints, and norm-specific requirements (often formally specified using deontic or declarative languages) that limit agent autonomy to ensure process compliance and goal alignment. Frames can exist at multiple levels (agent, process, organizational) and may be operational (prescribing specific sequences) or normative (specifying permitted/prohibited behaviors).

The separation of operational versus normative frames enables context-dependent autonomy: agents may optimize their own strategies within the boundaries of regulation, fostering distributed but process-compliant intelligence.

Explainability

Given the opacity of many modern AI models and the adaptive autonomy of agents, explainability is elevated to a first-class requirement in APM. Agents must provide explicit rationales for their behaviors—including traces, causal dependencies, and justifications for decisions—to ensure trust, debuggability, regulatory compliance, and traceable accountability. Explainability mechanisms must support both technical (audit logs, XAI techniques) and user-interpretable (dialogue, reports, recommendations) modalities, with explainable AI becoming integral to trustworthy agentic process execution.

Conversational Actionability

Agents must bridge conversational capabilities (interacting with humans and other agents using structured protocols or NL interfaces) with process enactment (effectuating actions, making decisions, or adapting workflows in real time). Conversational interfaces enable negotiation, decision review, and exception handling; actionability bridges this interaction to process-aligned execution, encompassing capabilities such as querying, recommending, executing, and creating process artifacts.

Self-Modification

APM agents require mechanisms for both short-term adaptation (ephemeral, instance-specific adjustments) and long-term evolution (structural changes to agent models or collective processes). Effective adaptation necessitates runtime monitoring, exception handling, feedback loops, and context-aware learning. Evolutionary changes may derive from systematic trend analysis, process mining, and causal inference over execution data, reinforcing successful adaptive strategies and mitigating recurrent failures.

Research Challenges and Open Problems

The manifesto identifies an extensive research agenda, structured according to capability domains and cross-cutting concerns:

  • Framed Autonomy: Open questions include formalizing the practitioner-understandable definition of "agent," practical frame elicitation (potentially employing LLMs for rule extraction), operationalizing frames in legacy or heterogeneous settings, and designing effective, learnable incentive/reinforcement structures for frame compliance.
  • Explainability: Technical challenges span the specification and dynamic adaptation of explanation preferences, integrating and extending XAI methods, generating actionable and situationally relevant explanations, managing explanation lifecycle, and ensuring causal soundness rather than mere correlation.
  • Conversational Actionability: Key problems include seamlessly integrating diverse human-agent and agent-agent interaction protocols (from NL to formal languages), mapping conversational inputs to process-aligned actions, enabling legacy systems to participate via agentic wrappers, and rigorously computing process-wide functional indicators and trade-offs between autonomy and control.
  • Self-Modification: Central issues involve the governance of self-modification (deciding deferral conditions to human oversight), evaluating modification outcomes, managing concurrent evolution/adaptation among process-linked agents, and incorporating bounded but effective continuous learning and adaptation mechanisms.
  • Cross-cutting: Migrating traditional BPM assets into agentic paradigms, developing APM-specific security and privacy frameworks (especially with explainability and self-modification), comprehensive benchmarking (addressing data contamination and generalization), legal liability and accountability frameworks, and engineering methodologies for systematic APM system design.

Implications and Future Research Directions

The delineation of APM marks a decisive evolution from static automation-centric BPM to highly dynamic, distributed, and process-aware multi-agent organizational systems. The paradigm foregrounds formal alignment between agentic autonomy and organizational control, requisite for deploying scalable, trustworthy, and explainable autonomous systems within enterprise backends.

Practically, APM architectures will enable adaptive, robust process execution in environments characterized by high volatility and ambiguity, increasing organizational agility and reducing dependency on brittle, human-dependent workflows. Theoretically, APM challenges the BPM and MAS communities to converge efforts in formal process specification, norm synthesis, XAI, and orchestration of distributed adaptive intelligence.

Future developments will depend on advances in:

  • Formal abstractions for frames, goals, and agentic alignment.
  • XAI techniques specialized for dynamic, multi-agent, process-aware systems.
  • Conversational AI tailored for process negotiation and exception management.
  • Continuous learning, causal process mining, and systematic adaptation/evolution in agent swarms.
  • Engineering methodologies integrating legacy system onboarding, comprehensive security, traceability, robust benchmarking, and applied legal frameworks.

Conclusion

"Agentic Business Process Management: A Research Manifesto" (2603.18916) offers a rigorous, conceptually grounded blueprint for the evolution of BPM toward process-aware, agent-driven systems. By formalizing the architecturally distinct roles of macro-level framing and micro-level agent execution—and enumerating the critical capabilities of framed autonomy, explainability, conversational actionability, and self-modification—the work systematically charts the research, engineering, and theoretical tasks required for APM realization. The manifesto calls for principled, interdisciplinary advances bridging BPM, AI, MAS, and software engineering, with an emphasis on reproducible software artifacts, empirical benchmarks, and the robust governance demanded by future AI-integrated organizational processes.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Explain it Like I'm 14

Agentic Business Process Management: A Simple Explanation

What is this paper about?

This paper introduces a new way for companies to manage work when smart “agents” (like AI programs, robots, or people using AI) do parts of the job on their own. The authors call this Agentic Business Process Management (APM). The big idea: let agents act autonomously, but keep them “process-aware” and inside clear boundaries so their decisions still follow the company’s rules, goals, and responsibilities.

What questions are the authors trying to answer?

The paper asks, in plain terms:

  • How can organizations safely use autonomous agents to do real business work?
  • What does it mean for agents to be “process-aware,” so their actions follow company goals and rules?
  • What basic abilities do these agents need to work well with people and other agents?
  • How should we design the overall system so everything stays aligned, explainable, and compliant?

How did the authors approach this?

This is a “manifesto,” not a lab experiment. Think of it as a roadmap or guide written by many experts after deep discussions:

  • They met at a research seminar and workshops, split into groups, and debated what’s needed for agent-driven work.
  • They combined ideas from business process management (BPM), AI, and multi‑agent systems (how multiple smart agents work together).
  • They proposed shared concepts, a high-level system design, and a list of key agent abilities and research challenges.

To explain a few terms with everyday analogies:

  • Business Process Management (BPM): Like a playbook for how work should be done in a company (who does what, in what order, to reach a goal).
  • Agent: A “co-worker” that can sense what’s happening, think about goals, and act—this can be a person, an AI bot, a robot, or regular software with some autonomy.
  • Process awareness: The agent doesn’t just act; it knows the company’s playbook (goals, rules, and constraints) and uses that to guide its choices.
  • Framing: The “guardrails” or game rules that limit what agents can do, so they stay safe, fair, and compliant.

What are the main ideas and findings?

The paper proposes a clear structure and four must‑have abilities.

  1. A two-level architecture
  • Macro level (the big picture): The APM system sets shared goals and guardrails (“the frame”) for all agents. It ensures agents’ personal goals line up with the organization’s goals.
  • Micro level (the workers): Individual agents—human or AI—run a loop of Perceive → Reason → Act. They use tools (like apps, sensors, or services), talk to each other, and make choices within the frame.
  1. The four key capabilities agents need To make APM work in practice, agents should have:
  • Framed autonomy: Freedom to make smart decisions, but within clear rules and constraints (the “frame”). The paper distinguishes:
    • Normative frames: “What’s allowed or required?” (like laws or company policies)
    • Operational frames: “How do we do it?” (step-by-step instructions)
  • Explainability: Agents should be able to explain why they did something. This builds trust, helps debugging, supports fairness, and shows compliance with laws (like the EU AI Act or GDPR).
  • Conversational actionability: Agents should communicate naturally and cooperate—negotiating, coordinating, and taking action through dialogue with people and other agents.
  • Self-modification: Agents should improve over time—learning from outcomes, updating plans, and adapting while still respecting the frame.
  1. Examples to make it concrete
  • Supplier onboarding: Buyer and supplier agents (some human, some AI) request and evaluate quotes, choose contracts, and ensure terms are met—always staying within the process rules.
  • Ride-sharing: A passenger agent books a ride, an AI agent allocates cars, a vehicle agent plans routes, and a payment agent handles billing—all coordinated under shared goals and constraints.
  1. Why this matters
  • Autonomy without chaos: Agents can act smartly and quickly, but stay aligned with rules and goals.
  • Safer AI in business: Explainability and framing reduce risks like breaking policies, being unfair, or making unsafe choices.
  • Better collaboration: Conversational agents help humans stay in the loop and adjust plans on the fly.
  • Continuous improvement: Self-modification lets systems get better over time without losing control.

What’s the impact if this works?

If companies adopt APM:

  • They could automate complex work more safely and responsibly, improving speed and quality while reducing errors and risk.
  • Teams would trust AI agents more because actions are explainable and auditable.
  • Organizations could handle changing laws, policies, and markets better by updating frames and letting agents adapt.
  • Researchers in BPM and AI would have a shared roadmap to build practical, compliant, and human-aware agent systems.

In short: The paper argues that the future of using AI at work isn’t just “more automation.” It’s “guided autonomy”—agents that think and act on their own, but within clear, explainable boundaries that keep them aligned with what the organization and society expect.

Knowledge Gaps

Knowledge gaps, limitations, and open questions

The paper is a forward-looking manifesto and intentionally leaves many aspects underspecified. The following list distills concrete gaps and open questions that future research should address to operationalize Agentic Business Process Management (APM):

  • Formal semantics for process awareness and framing: precise definitions of “process awareness,” “normative frame,” and “operational frame,” their compositional semantics, and mappings to existing BPM languages (e.g., BPMN, DECLARE, DMN).
  • Engineering patterns for framing: concrete mechanisms to realize framing (orchestration vs. shared memory vs. decentralized governance), assignment of frame responsibility, runtime enforcement, and monitoring.
  • Strategy synthesis under frames: algorithms for planning and decision-making with normative constraints in (i) single decision-maker, (ii) multiple decision-makers with local frames, and (iii) multiple decision-makers with process-level frames; support for partial observability and stochastic environments.
  • Projection and reconciliation of norms: methods to project global normative frames to local agent-level norms, detect and resolve conflicts, and establish soundness/completeness of such projections.
  • Verification and assurance: scalable static verification and runtime monitoring that agent behaviors conform to frames; conformance checking for agentic logs; techniques to cope with non-deterministic LLM-based decision modules.
  • Metrics for autonomy and alignment: operational measures of “degree of process awareness,” autonomy, and goal alignment at agent and system levels; KPIs to compare APM designs and configurations.
  • Reference architecture and interfaces: concrete APM reference architecture, component boundaries, and interface contracts for agents, tools, shared memory, and governance services; integration points with BPM suites.
  • Interoperability and standards: how APM integrates with and extends BPMN/DMN/CMMN, FIPA/ACL, MCP/ACP; whether new notations are needed for goals/frames and how to achieve cross-vendor interoperability.
  • Data and event models for observability: logging schemas connecting agent actions to process instances; extensions to process mining for multi-agent settings (e.g., agent system mining, conformance under norms); shared benchmarks and datasets.
  • Explainability design: when/what/how to explain at both agent and process levels; explanation formats that combine operational and normative rationales; cross-agent explanations and aggregation of rationales.
  • Explanation quality and governance: quantitative and qualitative measures of explanation usefulness, completeness, and faithfulness; triggers, recipients, escalation policies; trade-offs with autonomy and performance; regulatory adequacy (e.g., GDPR/AI Act).
  • Conversational actionability protocols: standards for negotiation, coordination, and human-in-the-loop control; grounding in shared ontologies; resilience to prompt injection, social engineering, and conversation drift.
  • Human–AI governance: roles, oversight models, and accountability (e.g., RACI for agents and humans); liability frameworks for agent decisions; organizational change management and practitioner training for APM design.
  • Self-modification boundaries and safeguards: policies specifying what agents may change (knowledge, goals, tools, policies), approval gates, rollback/versioning, and mandatory post-change verification and auditing.
  • Continual learning and concept drift: methods for agents to adapt without violating frames; synchronization of learned behaviors with process specifications; prevention and detection of specification drift.
  • Secure and safe tool use: capability modeling, least-privilege access control, sandboxing, and monitoring for tool invocation; detection/mitigation of tool misuse and adversarial inputs; robustness to external service failures.
  • Performance and scalability: quantifying overheads of framing, explainability, and monitoring; scaling to many agents/processes with acceptable latency and cost; performance–safety trade-offs.
  • Cyber-physical integration: extending APM to embodied/robotic agents; real-time safety constraints, verification, and latency in physical environments.
  • Cross-organizational APM: frames spanning organizational boundaries; choreography under contracts and regulations; data sovereignty, privacy, and secure shared memories in inter-organizational processes.
  • Migration pathways: methods and tools to refactor existing BPM/RPA assets into frames and agent capabilities; incremental adoption strategies, compatibility layers, and ROI analysis.
  • Decision core design: to what extent LLMs can serve as agent reasoning modules; design of hybrid architectures (symbolic planners, constraints, program synthesis + LLMs); calibration, testing, and fallback strategies.
  • Pre-deployment validation: simulation and digital twins to test frames and agent interactions; scenario-based risk assessment, what-if analysis, and stress testing.
  • Empirical validation: real-world case studies and controlled experiments quantifying benefits, risks, and failure modes of APM; user studies on trust, usability of explanations, and governance efficacy; ablation studies of the four capabilities’ contributions.

Glossary

  • Agent Communication Protocol (ACP): A standardized protocol for agent-to-agent communication. "Examples include Model Context Protocol (MCP), Agent Communication Protocol (ACP), Resource Description Framework (RDF), and Knowledge Query and Manipulation Language (KQML)."
  • Agent governance: The policies and mechanisms for regulating and overseeing autonomous agents’ behavior. "Agent governance has been a well-established line of research for several decades"
  • Agent-oriented modeling: Modeling approaches that center on agents, their goals, and interactions (e.g., i*, Tropos). "agent-oriented modeling frameworks like i* and Tropos"
  • agent system mining: Process mining techniques focused on extracting and analyzing agent behaviors and interactions. "the emerging research directions of agent system mining"
  • agentic AI: AI approaches that frame software in terms of autonomous agents, often leveraging LLMs. "efforts to achieve greater autonomy in software systems often attempt to make use of LLMs and are summarized under the umbrella term of agentic AI"
  • agentic AI process observability: Techniques and tools to observe and analyze processes executed by agentic AI. "as well as of agentic AI process observability"
  • Agentic Business Process Management (APM): An extension of BPM focused on governing autonomous, process-aware agents executing processes. "This paper presents a manifesto that articulates the conceptual foundations of Agentic Business Process Management (APM)"
  • agentic system: A software system composed of goal-driven agents that sense, reason, and act. "we introduce the concept of an agentic system as a collection of (one or more) individual goal-driven agents"
  • AI Act: The European Union’s regulatory framework governing AI systems. "regulatory frameworks, such as the EU's GDPR and AI Act"
  • AI-augmented BPM systems: BPM systems enhanced with AI capabilities but not centered on agents as primary executors. "This marks a paradigm shift from previously introduced AI-augmented BPM systems"
  • BDI (Belief–Desire–Intention): An agent architecture modeling mental attitudes to drive decision-making. "classic AI constructs such as BDI and FIPA"
  • Business Process Management (BPM): The discipline of managing coordinated activities to achieve organizational goals. "Business Process Management (BPM) literature and practice"
  • BPMN (Business Process Model and Notation): A standard for modeling business processes with graphical notation. "classical process specification languages, such as BPMN and DECLARE"
  • DECLARE: A declarative process modeling language specifying constraints instead of fixed flows. "classical process specification languages, such as BPMN and DECLARE"
  • deontic requirements: Normative stipulations (obligations, permissions, prohibitions) that constrain behavior. "frames are normative: they specify deontic requirements to the process"
  • DOLCE (Descriptive Ontology for Linguistic and Cognitive Engineering): A foundational ontology used to structure domain knowledge. "Drawing on foundational ontologies like DOLCE and GFO"
  • eXplainable AI (XAI): Methods that make AI decisions understandable by providing explanations. "which explanation mechanisms---so-called eXplainable AI (XAI) techniques---may be employed"
  • Embodied agents: Agents realized in physical form (robots) that act in the physical world. "(Physically) Embodied agents (a.k.a robots)"
  • FIPA (Foundation for Intelligent Physical Agents): Standards and specifications for agent technologies. "classic AI constructs such as BDI and FIPA"
  • foundation models: Large, pre-trained AI models adaptable to many tasks (e.g., LLMs). "foundation models---such as LLMs---organizations increasingly adopt"
  • four-eyes approval policies: Governance rules requiring approval by at least two independent actors. "regarding four-eyes approval policies"
  • framed autonomy: Autonomy that is constrained and aligned through explicit frames (rules and goals). "four key capabilities that such APM agents must support: framed autonomy, explainability, conversational actionability, and self-modification"
  • GDPR (General Data Protection Regulation): The EU’s data protection and privacy regulation. "regulatory frameworks, such as the EU's GDPR and AI Act"
  • GFO (General Formal Ontology): A foundational ontology for representing general categories and relations. "foundational ontologies like DOLCE and GFO"
  • i*: A goal- and dependency-oriented modeling framework for requirements and organizations. "agent-oriented modeling frameworks like i* and Tropos"
  • KQML (Knowledge Query and Manipulation Language): A language and protocol for exchanging information and knowledge among agents. "Resource Description Framework (RDF), and Knowledge Query and Manipulation Language (KQML)."
  • LLMs: Generative AI models trained on large corpora to process and produce language. "foundation models---such as LLMs---"
  • Mental model theory of reasoning: A cognitive theory explaining how humans reason using mental representations. "Mental model theory of reasoning"
  • Model Context Protocol (MCP): A protocol to standardize how models are provided with and operate over context. "Examples include Model Context Protocol (MCP), Agent Communication Protocol (ACP), Resource Description Framework (RDF), and Knowledge Query and Manipulation Language (KQML)."
  • Multi-Agent Systems (MAS): Systems composed of multiple interacting agents with possibly different goals. "promoting the notion of agents and Multi-Agent Systems (MAS) as the fundamental abstractions of software systems"
  • normative frame: A frame capturing norms (obligations, permissions, prohibitions) that constrain agent behavior. "we call the two frames normative frame and operational frame"
  • normative MAS: MAS that regulate agent behavior through formal norms to ensure compliance and social order. "the AI subfield of normative MAS, which focuses on regulating autonomous agent behavior through the specification of norms"
  • operational frame: A frame that prescribes how processes should be executed operationally (e.g., steps, data handling). "we call the two frames normative frame and operational frame"
  • orchestration: Centralized coordination of activities or agents to realize a process. "for example, by assigning collective responsibility to one or more agents, through orchestration, or via a shared memory"
  • Perceive--Reason--Act loop: The continual cycle in which an agent senses, decides, and acts. "at the micro level each process-aware agent runs a Perceive--Reason--Act loop over framed knowledge and goals"
  • process awareness: An agent’s awareness of process models, constraints, and goals, used to align its behavior. "We define process awareness as the assurance that agents’ inner workings conform to organizational processes and adhere to their operational constraints, regulations, and goals."
  • process choreographies: Specifications of interactions among independent process participants without central control. "notably, work on process choreographies"
  • process mining: Data-driven techniques to discover, monitor, and improve real processes from event logs. "within process mining"
  • RDF (Resource Description Framework): A W3C framework for representing information about resources on the Web. "Agent Communication Protocol (ACP), Resource Description Framework (RDF), and Knowledge Query and Manipulation Language (KQML)."
  • RPA (Robotic Process Automation): Automation technology using software robots to mimic deterministic human tasks. "Unlike traditional BPM or Robotic Process Automation (RPA) systems that follow rigid, predefined rules and workflows"
  • shared memory: A shared state or storage used by agents to coordinate and maintain common context. "through orchestration, or via a shared memory (e.g., as part of the frame) that allows agents to progress along a mutual plan."
  • socio-technical system: A system comprising both social (human) and technical (technological) elements. "An APM system is an agentic socio-technical system, jointly realized by a collection of agents"
  • technical debt: The long-term maintenance cost incurred by expedient or suboptimal design decisions. "leading to system integration and maintenance challenges and, ultimately, to technical debt"
  • Tropos: An agent-oriented software development methodology emphasizing goals and stakeholders. "agent-oriented modeling frameworks like i* and Tropos"

Practical Applications

Immediate Applications

Below are deployable use cases that apply the paper’s APM principles (framed autonomy, explainability, conversational actionability) using today’s BPM, AI, and integration tooling.

  • Procurement and supplier onboarding with framed AI agents
    • Sectors: manufacturing, retail, pharma
    • What it looks like: Buyer and supplier agents issue/respond to RFQs, validate documents, negotiate terms, and route approvals under a normative frame (e.g., spend limits, four‑eyes approval, due diligence).
    • Tools/products/workflows: BPM suites (BPMN/DMN/DECLARE) + policy/constraints engine, LLM-based agents with tool use (MCP/ACP), document intelligence, e-signature, ERP integration.
    • Assumptions/dependencies: Clean master data; machine-readable policies; role-based access control (RBAC); human-in-the-loop for exceptions.
  • Accounts payable and order-to-cash exception handling with guardrails
    • Sectors: finance, shared services, SaaS
    • What it looks like: Agents match POs/invoices, chase missing goods receipts, explain mismatches, and propose resolution plans, bounded by frames (e.g., SOX controls, segregation of duties).
    • Tools/products/workflows: Process mining to detect deviations; AI agents orchestrated by BPM; explainability console with action traces; ERP/ITSM connectors.
    • Assumptions/dependencies: Well-defined approval hierarchies; audit logging; API coverage for core systems.
  • Customer service triage and resolution under normative frames
    • Sectors: telecom, software, e-commerce
    • What it looks like: Agents classify tickets, propose next best actions, negotiate SLAs with upstream/downstream teams, and escalate per policy; provide explanations for routing/decisions.
    • Tools/products/workflows: ITSM (ServiceNow/Jira) + LLM agents with tool use; conversation-to-case frames; SLA policy engine; chatops interfaces.
    • Assumptions/dependencies: High-quality taxonomy; documented SLAs and escalation policies; conversation safety filters.
  • Healthcare administrative coordination (non-clinical)
    • Sectors: healthcare, payers/providers
    • What it looks like: Agents schedule appointments, handle prior-authorization paperwork, manage discharge logistics under compliance frames (HIPAA/GDPR), and produce explanations for denials.
    • Tools/products/workflows: BPM + EHR integration; document extraction; agent guardrails for PHI; audit trails for compliance.
    • Assumptions/dependencies: Clear data-sharing agreements; strong identity and privacy controls; human oversight for high-risk cases.
  • Fair and compliant HR recruiting workflows
    • Sectors: all with high-volume hiring
    • What it looks like: Agents screen applicants, schedule interviews, and communicate feedback constrained by fairness and privacy frames (e.g., anti-discrimination, data retention).
    • Tools/products/workflows: ATS integration; XAI to justify screening decisions; policy-as-code for fairness constraints; candidate-facing conversational interfaces.
    • Assumptions/dependencies: Bias-mitigation policies codified; auditable logs; consent and data minimization in place.
  • Framed dispatch and routing in mobility/logistics
    • Sectors: ride-hailing, field service, last-mile delivery
    • What it looks like: AI agents allocate vehicles/technicians and optimize routes while staying within safety, zoning, and service-level frames; provide rationale for assignments.
    • Tools/products/workflows: Optimization engines + agent layer; telemetry; explainable dispatch dashboard; geofencing policies.
    • Assumptions/dependencies: Reliable location data; regulatory constraints encoded as norms; override mechanisms.
  • “Process copilot” for managers and operators (conversational actionability)
    • Sectors: cross-industry
    • What it looks like: Natural language interactions to query status (“Why is this case stuck?”), trigger framed actions (“Start rework path”), and modify thresholds within authority.
    • Tools/products/workflows: Chat interfaces backed by process knowledge graphs and BPM APIs; role-aware guardrails; change request workflows.
    • Assumptions/dependencies: Up-to-date process models; robust authorization; change impact checks.
  • Agentic observability and explainability for existing automations
    • Sectors: any with RPA/BPM
    • What it looks like: Wrap current automations with agent-style traces and explanations (who/what/why), enabling governance dashboards and faster incident response.
    • Tools/products/workflows: Process mining, event collection, XAI wrappers; “explain-why” endpoints exposed by agents and services.
    • Assumptions/dependencies: Sufficient event coverage; standardized logging schemas; data retention policies.
  • B2B negotiation agents under shared frames
    • Sectors: manufacturing, wholesale, insurance
    • What it looks like: Buyer/seller agents negotiate terms under agreed normative frames (payment windows, risk tolerances, compliance clauses) and maintain auditable transcripts.
    • Tools/products/workflows: Agent communication protocols (MCP/ACP); contract clause libraries; e-invoicing/EDI connectors; negotiation playbooks as norms.
    • Assumptions/dependencies: Legal alignment on machine-interpretable norms; dispute resolution mechanisms; counterparty risk checks.
  • Research and teaching sandboxes for APM
    • Sectors: academia, industry labs
    • What it looks like: Benchmarking framed autonomy scenarios; synthetic multi-agent processes with varying observability, frames, and explainability demands.
    • Tools/products/workflows: Open-source APM reference architecture; synthetic logs; evaluation harnesses for alignment and compliance.
    • Assumptions/dependencies: Shared datasets/ontologies; reproducible baselines; ethics approvals for user studies.

Long-Term Applications

The following use cases depend on advances in multi-agent reasoning, formal verification, standards, safety, and scalable integration. They reflect the paper’s vision, including self-modification.

  • End-to-end autonomous process execution across enterprises
    • Sectors: supply chain, finance, pharma
    • What it looks like: Multi-agent execution of procure-to-pay/order-to-cash with dynamic, cross-firm frames (contracts, regulations), minimal human intervention, and continuous explainability.
    • Dependencies: Interoperable normative standards; cross-organizational identity and trust; robust verification under partial observability; dispute resolution frameworks.
  • Self-modifying processes with safe learning and governance
    • Sectors: cross-industry
    • What it looks like: Agents propose and implement process/frame updates (e.g., new approval path) based on outcomes; changes are verified, simulated, and rolled out with rollback plans.
    • Dependencies: Formal change control, causal/impact analysis, safe RL/continual learning, “frame diff” tooling, and certification workflows.
  • Machine-readable regulation to frame agents (RegTech at scale)
    • Sectors: finance, healthcare, energy, public sector
    • What it looks like: Laws/policies encoded as normative frames consumed by APM systems for continuous compliance; regulators audit via explainable traces.
    • Dependencies: Regulatory standards for computable norms; co-regulatory sandboxes; liability and auditability schemas; legal/technical harmonization.
  • Autonomous compliance and audit for high-risk domains
    • Sectors: banking, insurance, clinical operations
    • What it looks like: Agents continuously reconcile actions against dynamic norms, flag risks preemptively, and generate legally sound explanations and evidence packages.
    • Dependencies: Assurance cases; certified XAI; evidence provenance; privacy-preserving audit tech.
  • Human–AI mixed-initiative process redesign
    • Sectors: cross-industry
    • What it looks like: Agents discover bottlenecks from logs, simulate alternatives, propose redesigns, and coordinate adoption—within strategic frames and constraints.
    • Dependencies: Reliable process mining across agents; simulation fidelity; organizational change management; incentive alignment.
  • Safety-critical embodied APM on factory floors
    • Sectors: manufacturing, warehousing, construction
    • What it looks like: Robots and human agents coordinate tasks under safety and productivity frames, with formal verification of norms and runtime monitoring.
    • Dependencies: Verified controllers; certified human-robot interaction norms; low-latency observability; fallback/stop mechanisms.
  • Personalized education orchestrated by framed agents
    • Sectors: education, edtech
    • What it looks like: Agents plan learning paths aligned with curricular frames, ensure fairness (e.g., accessibility), and explain progression decisions to students/teachers.
    • Dependencies: Trusted student models; fairness and privacy frames; content quality assurance; teacher oversight.
  • Smart grid and energy markets with framed autonomy
    • Sectors: energy, utilities
    • What it looks like: Agents orchestrate demand response, storage, and trading within stability and regulatory frames; self-adapt to grid events with explainable actions.
    • Dependencies: Real-time telemetry; interoperable market norms; safety constraints; resilience against cascading failures.
  • Framed algorithmic trading and treasury operations
    • Sectors: capital markets, corporate finance
    • What it looks like: Agents execute trades and liquidity moves within risk, compliance, and ethical frames; adapt strategies with self-modification under strict oversight.
    • Dependencies: Verified risk constraints; market abuse detection; auditable strategies; regulator-approved explainability.
  • Public administration and permitting with APM
    • Sectors: government
    • What it looks like: Agents process permits/benefits under legal frames, coordinate across departments, and provide clear explanations to citizens.
    • Dependencies: Digitized statutes; inter-agency data sharing; transparency mandates; redress mechanisms.
  • Standards and infrastructure for agent interoperability
    • Sectors: software, platforms
    • What it looks like: Broad adoption of protocols (MCP/ACP), shared ontologies, and secure tool-access policies enabling plug-and-play process-aware agents.
    • Dependencies: Industry consortia; security and isolation frameworks; certification of agent capabilities; economic incentives for standardization.
  • Security-by-design for agent tool use and isolation
    • Sectors: all
    • What it looks like: Capability-based permissions, sandboxed tool execution, and continuous policy enforcement embedded in APM frames.
    • Dependencies: Strong identity/attestation for agents; fine-grained policy engines; runtime monitoring and kill-switches.

Notes on feasibility across applications:

  • Many immediate applications rely on existing BPM/RPA stacks, LLM-based agents with tool use, policy-as-code engines, and strong human-in-the-loop controls.
  • Long-term applications require advances in formal normative reasoning, multi-agent coordination under uncertainty, computable regulation, certified explainability, interoperability standards, and robust governance frameworks.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 71 likes about this paper.