Process mining for AI opportunities starts with an uncomfortable observation: most enterprise AI pilots were chosen in a room. Someone drew a 2×2 matrix on a whiteboard, senior leaders debated which workflows sounded exciting, and three months later a team is building something the system logs never asked for. The data already knew which processes were painful, repetitive, and ripe for automation. Nobody looked.
Process mining — practiced by tools like Celonis, Apromore, Disco, and IBM Process Mining — extracts event logs from systems of record (SAP, Oracle, ServiceNow, Salesforce, Jira) and reconstructs every actual path your processes take[5]. Not the SOP from 2019. Not the BPMN diagram the consultant drew. The literal sequence of timestamped events from your production systems. What comes out is a ranked list of workflows by volume, cycle time, bottleneck severity, and variant count — the raw material for an evidence-based AI use case pipeline.
The market for this category is growing fast. Process mining software is projected to grow from roughly $850M in 2026 toward $2B by 2031, with Celonis claiming 60% market share and consistently landing as the category leader in analyst assessments[1]. But most adoption is still RPA-flavored — using mining to find rule-based automation candidates. The LLM-era question is different: which mined workflows are good candidates for language model automation versus which ones need rule execution versus which ones aren't worth automating at all?
Why Workshop-Driven Selection Produces Bad Picks
Three structural reasons why the room always picks the wrong workflows — and why process data cuts through all three
Workshop-driven AI use case selection fails predictably, and it fails for the same three reasons across industries.
The loudest voice wins. Every executive prioritization session has someone senior enough to override the analysis. They've seen a vendor demo, they have a pet project, or they just find the idea of "AI for contract review" more exciting than "AI for invoice routing." The team builds the interesting idea, not the high-ROI one.
Nobody wants to admit which workflow is actually painful. Ask a room of VPs which process is the most broken and they'll describe a process in someone else's department. The order-to-cash exception handling that costs the finance team 40 hours a week never comes up because the finance VP doesn't want to air dirty laundry in front of the CIO. Process mining surfaces the pain from system data — not from organizational politics.
The interesting idea always beats the measurable one. "AI for customer sentiment analysis" sounds innovative. "AI to reduce the 8% exception rate in supplier onboarding" sounds unglamorous. Both can have identical ROI. The second one has a cleaner feedback loop, less hallucination risk, and a more straightforward success metric. Workshop groups consistently pick the first one. Mining-based approaches let the bottleneck data make the initial ranking before any human applies preferences.
Process Mining in One Page (For the AI Reader)
ETL + clustering + visualization on event logs — and why the output changes what AI selection looks like
If your background is ML, not BPM, here's what process mining actually does. Every enterprise system of record — SAP, Oracle, ServiceNow, Salesforce, Jira — emits an event log: a timestamped record of which case (invoice, ticket, contract, employee record) hit which step, when, and in what sequence. Process mining tools ingest those logs and run clustering and graph algorithms over the case sequences to reconstruct a process graph.
The output isn't a theoretical BPMN diagram. It's a map of every distinct path actually taken, annotated with frequency and duration. A workflow your SOP says has two paths might have 47 variants in reality — each deviation representing a data quality issue, an edge case someone built a workaround for, or a policy exception that became the default. The variant count is one of the most diagnostic signals in the whole analysis: it tells you whether you have a stable process (low variants, safe for RPA) or a judgment-intensive one (high variants, candidate for LLM treatment).
Process mining tools also compute conformance checking (how often the actual process matches the intended one), bottleneck analysis (which steps accumulate wait time), and root cause attribution (which attributes correlate with slow cases). This gives you, for every workflow in your organization's event log, a quantified profile: how big it is, how slow it is, how variable it is, and what's causing the pain. That profile is the input to an AI suitability scoring model — not a whiteboard session.[5]
SOP documents written in 2019 and never updated
Happy-path diagrams showing 1-2 process variants
Estimated cycle times from interviews with process owners
Bottlenecks identified by whoever complained loudest last quarter
BPMN diagrams that no longer match the system configuration
Log-derived event sequences from actual production systems
Every variant captured — often 20–80 distinct paths through the same process
Real durations from timestamped event data, not estimates
Bottleneck ranking by total accumulated wait time across all cases
Conformance scores showing where the actual process deviates from intent
The AI Suitability Score: Volume × Variance × Structure
A three-axis framework for ranking mined processes against each other — and for deciding whether a workflow needs an LLM, an RPA bot, or neither
Once you have the mined process data, you need a scoring model that translates it into a ranked candidate list. The AI Suitability Score is a three-axis product: Volume × Variance × Structure. Score each axis 1–5, multiply them, and you have a comparable number for every workflow in your inventory.
Volume measures ROI ceiling. A workflow that processes 50,000 cases per year at 25 minutes per case represents 20,000+ hours of annual labor. A workflow that runs 50 times a year is a rounding error. Volume is the non-negotiable axis — without sufficient throughput, no automation investment pays back. Score it: 1 = under 1,000 cases/year, 3 = 10,000–50,000 cases/year, 5 = 100,000+ cases/year.
Variance is the LLM's competitive advantage over RPA. RPA breaks on variance — every new path through a process requires a new rule. The standard rule of thumb is that a process with more than 10–15 variants is a poor RPA candidate. LLMs handle variance differently: they read the case, interpret the context, and apply judgment to the anomaly rather than failing on it. A high-variant process (scoring 4–5 on this axis) is exactly where you want a language model, not a bot. Score it: 1 = 1-3 variants (pure RPA territory), 3 = 10-20 variants (borderline), 5 = 30+ variants (LLM territory).
Structure measures input parsability. A fully structured workflow — all inputs are database fields, all decisions are rule-based — doesn't need an LLM. RPA or a deterministic rule engine handles it more cheaply. When inputs include unstructured text (emails, PDFs, free-form comments, contracts), audio, or require contextual judgment (is this expense claim within policy?), the calculus flips. Score it: 1 = fully structured inputs and rule-based decisions, 3 = mixed structured and unstructured, 5 = primarily unstructured inputs or judgment-required decisions.
The scoring model produces a ranked list. High-scoring workflows (above 40 out of 125) are genuine AI candidates — high enough volume to justify investment, high enough variance that RPA will fail, unstructured enough that LLMs outperform rule engines. Mid-range workflows (15–40) warrant a tool-selection conversation: some are clean RPA candidates, some need better data before they're automatable, some belong on a roadmap for later. Low-scoring workflows below 15 are usually best left alone — the ROI ceiling doesn't justify the build cost.
| Workflow | Volume | Variance | Structure | Score | Better Tool |
|---|---|---|---|---|---|
| Invoice approval (3-way match exceptions) | 5 | 3 | 3 | 45 | LLM |
| Customer support ticket triage | 5 | 5 | 5 | 125 | LLM |
| Contract redlining (standard templates) | 3 | 4 | 5 | 60 | LLM |
| Expense report validation | 4 | 2 | 2 | 16 | RPA |
| Employee onboarding (system provisioning) | 3 | 2 | 1 | 6 | RPA or neither |
| Supplier onboarding (documentation review) | 3 | 4 | 4 | 48 | LLM |
| Insurance claims adjudication (complex) | 4 | 5 | 5 | 100 | LLM |
| IT ticket routing (category assignment) | 5 | 3 | 4 | 60 | LLM |
The Discovery Pipeline: From Logs to Ranked Candidates
A five-step sequence that takes you from raw system event data to a prioritized shortlist of AI automation candidates
The pipeline isn't complicated, but the sequencing matters. Organizations that skip straight to AI suitability scoring without first running variant analysis and bottleneck profiling end up scoring the wrong attributes — what they assumed about the process rather than what the log data shows.
Step 1: Source system selection. Not every system of record produces clean event logs. SAP and Oracle are the richest sources for finance and procurement workflows. ServiceNow surfaces IT and HR processes. Salesforce covers sales and customer success workflows. Jira and similar tools capture engineering and operations processes. The practical constraint is usually connector availability — Celonis has hundreds of pre-built extractors; Apromore and open-source tools require more manual ETL work. Start with the two systems that cover your highest-volume operational workflows, not the ones that are easiest to connect.
Step 2: Variant analysis. Once the log is loaded, the mining tool constructs every distinct case sequence and ranks them by frequency. This is the moment of truth. The variant map tells you whether what you thought was a single workflow is actually six workflows under one name. High-variant processes (30+) immediately flag as LLM territory. Low-variant, high-frequency processes are clean RPA candidates. Processes with high variant counts but low frequency might be worth decomposing: the top 3 variants might be automatable even if the long tail is messy.
Step 3: Bottleneck identification. The mining tool shows where cases accumulate wait time — which step holds cases for the longest aggregate hours across the year. This is different from cycle time. A step that takes 2 minutes per case but processes 100,000 cases per year at 60% yield is a massive bottleneck even though individual cases move quickly through it. Automate the bottleneck, not the step that's technically interesting. The ROI calculation is straightforward: hours accumulated × average hourly cost × fraction automatable.
Step 4: AI suitability overlay. Apply the Volume × Variance × Structure scoring model to the top 20 bottlenecks by accumulated wait time. This produces your ranked candidate list. Anything scoring above 40 is worth a pre-mortem conversation. Anything above 70 is worth a funded pilot.
Step 5: Pre-mortem and commit. Before committing to the top 3 candidates, run a structured pre-mortem against each one (covered in the next section). The candidates that survive the pre-mortem become your committed pilots. The ones that don't go back into the pipeline for a future cycle.
The Tool Landscape (Honestly Assessed)
Six process mining platforms evaluated by fit, cost, and what they actually integrate with
The process mining software market is dominated by Celonis, which claims roughly 60% market share and consistently leads analyst rankings[1][4]. But market leadership doesn't mean best fit for your stack. The honest assessment depends on three questions: what are your primary source systems, how much technical capacity do you have for setup, and what's your budget horizon?
The enterprise space consolidated significantly in 2021 when SAP acquired Signavio, UiPath acquired ProcessGold, and IBM acquired myInvenio — all within months of each other. That consolidation was a signal: hyperscalers recognized that process discovery was a strategic moat for their automation suites[2]. The consequence is that your choice of mining tool is now partly a bet on your broader vendor ecosystem.
Tools worth evaluating in 2026
- ✓
Celonis — Market leader at 60% share, purpose-built for large enterprise transformation. Strongest SAP and Salesforce connectors, hundreds of pre-built extractors. Expensive at enterprise tier; requires PQL literacy; the right choice when SAP is your core system of record.
- ✓
Apromore — Open-core with a Community Edition (free) and an Enterprise Edition. Academic roots at University of Melbourne; strong on non-SAP source systems. Acquired by Salesforce, increasingly chosen by mid-market. Best pick for organizations that want control over the underlying model without a Celonis budget.
- ✓
Disco (Fluxicon) — Lightweight, single-analyst desktop tool. Best for quick process discovery sprints where you want to understand one workflow without a platform deployment. Not a full enterprise solution but fast to start with clean XES or CSV exports.
- ✓
UiPath Process Mining (formerly ProcessGold) — Deeply integrated with UiPath's RPA suite. If you're already on UiPath for automation execution, the discovery-to-deployment loop is tighter here than anywhere else. RPA-flavored by heritage; less suited for LLM-focused discovery.
- ✓
IBM Process Mining (formerly myInvenio) — Embedded in IBM Cloud Pak for Business Automation, with watsonx integration. Best for organizations on IBM infrastructure or z/OS environments. Discovery and automation execution stay inside the IBM stack.
- ✓
SAP Signavio — SAP-native process intelligence, tightly integrated with the SAP Business Technology Platform. The obvious choice if SAP is your source of truth and you want minimal integration friction. Less competitive outside the SAP ecosystem.
The Pre-Mortem: What Will Block This Automation
High suitability scores don't guarantee successful automations — run the pre-mortem before committing budget
The AI Suitability Score tells you whether a workflow is technically worth automating. The pre-mortem tells you whether your organization can actually automate it. These are different questions, and conflating them is how teams spend six months building pilots that never deploy.
The pre-mortem has five blockers to check. Data access: does your team have the permissions and API access to the systems involved, or will you spend three months waiting for InfoSec approvals? Change management: who owns this process today, and are they a sponsor or a blocker? Audit and regulatory: does this workflow have compliance constraints that require human sign-off at specific steps, and have you confirmed with legal that automation is permissible? Integration cost: what is the actual API work to connect the upstream source and downstream action systems, and does it fit in your pilot budget? Ownership dispute: does this workflow span multiple departments, and is there a clear single owner for the automation program? If two departments share ownership, the pilot will stall in cross-team approval loops.
Pre-mortem checklist for the top-ranked candidates
- ✓
Data access confirmed: system API credentials and read permissions secured for all source systems in the workflow
- ✓
Process owner identified and enrolled as a sponsor — not just informed, actively co-owning the pilot outcome
- ✓
Legal and compliance review completed: audit trail requirements documented, automation permissibility confirmed
- ✓
Integration cost estimated: upstream triggers and downstream actions mapped, not just the LLM inference step
- ✓
Change management plan sketched: who needs to change their workflow, when are they told, who trains them
- ✓
Success metric agreed: one primary metric (cycle time reduction, exception rate, cost per case) that all stakeholders accept as the decision criterion for scaling or stopping
Task Mining: When Logs Aren't Enough
When the workflow lives in screens instead of databases, you need a different discovery method
Process mining works on system-of-record event logs. But a significant share of enterprise work happens in surfaces that don't emit structured logs: Excel spreadsheets, email threads, browser-based internal tools, legacy desktop applications. When an analyst copies data from an ERP export into a spreadsheet, reconciles it against a PDF invoice, pastes the result into a web form, and emails the approver — none of those steps are in your SAP event log. They're invisible to process mining.
Task mining fills this gap by recording screen activity (keystrokes, mouse movements, application context) and using computer vision and NLP — increasingly LLM-based — to reconstruct the actual task sequence[3]. Tools like UiPath Process Mining's task capture module, Celonis Task Mining, and Microsoft Power Automate Process Mining all offer desktop recording modes. The output is similar to process mining: a variant map of how analysts actually execute screen-level tasks, with frequency and duration data.
The practical limitation is consent and privacy. Screen recording requires explicit employee consent and careful data governance — especially in regulated industries. The setup overhead is higher than connector-based process mining, and the data quality depends on recording duration and coverage. For knowledge worker workflows that have high strategic value but no system-of-record footprint, task mining is the right tool. For everything with clean event logs, stick with traditional process mining.
Anti-Patterns That Waste the Mining Budget
Five ways organizations invest in process mining without getting actionable AI candidates out the other side
Mining Without a Stakeholder Sponsor
Process mining surfaces uncomfortable truths about which processes are broken and who owns them. Without a C-suite sponsor who has explicitly committed to acting on the findings, the output becomes a beautiful process map that nobody uses to make a decision. Secure the sponsor before the tool contract.
The Complete Map Trap
Some teams insist on mapping every operational process before selecting any automation candidates. This produces a six-month enterprise-wide process inventory and no pilots. Mine two source systems, score the top 20 candidates, pick three. The remaining processes will still be there in the next 90-day cycle.
Ignoring the Bottleneck That Isn't a Workflow Problem
Process mining will faithfully show you that cases accumulate for 48 hours at a specific approval step. What it won't tell you is that the 48-hour wait is because one approver is overloaded and the backlog is a staffing decision, not an automation opportunity. Always ask whether a bottleneck has a human solution before building a technical one.
The Vanity Variant Count
High variant counts are a signal, not a verdict. Thirty process variants might mean 28 of them account for 2% of cases — the high-frequency variants are actually stable and automatable. Always segment by variant frequency before deciding a process is too variable for RPA. The long tail of variants is often noise, not signal.
Treating Mining as a One-Time Project
Process mining is most valuable as a continuous intelligence layer, not a one-off discovery exercise. Processes change: systems are updated, policies shift, volume spikes during seasonal periods. Organizations that run one mining sprint, select three automations, and then cancel the license find their automation portfolio drifting out of sync with how processes actually operate.
What This Looks Like in Your First 90 Days
A concrete sequence from zero to committed pilots — without boiling the ocean
- 1
Pick two source systems and connect them (Days 1–30)
Choose the two systems of record with the highest operational process volume — typically SAP or Oracle for finance/procurement, and ServiceNow or Jira for IT/operations. Avoid starting with CRM systems; their data quality is lower and the processes are harder to scope. Stand up the mining tool (a Disco trial works for a proof-of-concept sprint; Celonis or Apromore for a production deployment), extract 6–12 months of event log data, and get the first process graph rendered.
- 2
Run the 30-day mining proof (Days 15–45)
This is a focused analysis sprint, not a full process inventory. The goal is to score the top 20 bottlenecks from the two source systems using the AI Suitability model. Assign one process analyst (or a data-literate ops manager) to own the sprint. Output is a scored table of 20 candidates with notes on the primary blocking factors for each.
- 3
Score the top 10 candidates and run pre-mortems on the top 5 (Days 30–60)
Take the top 10 by AI Suitability Score into a structured pre-mortem session with the relevant process owners. Most organizations find that 2–3 of the top 10 are technically strong but organizationally blocked (data access, regulatory constraints, ownership disputes). Remove the blocked ones. The survivors become your pilot shortlist.
- 4
Commit to the top 3 with funding and timelines (Days 60–90)
Commit to three pilots with explicit funding, timelines, and success criteria. Structure them as the first article in this companion series recommends: one low-risk/high-signal, one medium-risk/high-value, one exploratory. Each should test something different about your organization's ability to execute automation. Set a go/no-go review date for each pilot at 60 days post-launch.
We had been running quarterly AI use case workshops for two years and had launched eleven pilots. Seven never reached production. When we finally ran a process mining sprint against our SAP and ServiceNow logs, the top three candidates by suitability score were workflows we had never discussed in a single workshop. They were too unglamorous to come up in a room full of VPs. All three are now in production.
Common Questions
The objections that come up in every process mining conversation
Do we need Celonis specifically, or can we DIY this?
You don't need Celonis. For a proof-of-concept sprint, Disco (Fluxicon) is free for limited datasets and produces variant maps and bottleneck analysis good enough to score your first 10 candidates. Apromore's Community Edition is open-source and covers the full process mining feature set. Celonis makes sense if you're deploying enterprise-wide with hundreds of processes and need pre-built SAP connectors, a monitoring layer, and commercial support. If you're doing a 90-day discovery sprint with one analyst, start with a lighter tool and graduate to a platform vendor after you've validated the approach works in your org.
What if our processes don't have clean event logs?
Most processes have better log data than their owners think. The real question is whether you have access to it and whether it's in extractable form. SAP stores event data in table structures that Celonis and similar tools know how to read. ServiceNow has native event logs. Even if your system doesn't have a native connector, any table with a case ID, an activity or status field, and a timestamp is enough to build an event log. The practical blocker is usually InfoSec permissions, not data absence. Start by mapping which of your systems of record have case-level timestamped status tables — you'll find more clean log data than you expect.
How do we handle workflows that span systems we don't own?
Cross-system workflows are common and they're harder to mine cleanly. The standard approach is to pick the system-of-record that holds the case for the longest portion of the workflow cycle time — that's where the bottleneck measurement will be most accurate. For handoff points between systems, you need a shared case identifier that persists across both systems (an invoice number, a ticket ID, a contract ID). Without a shared key, the cross-system join is unreliable. If you can't join the logs reliably, mine each system separately and treat the handoff itself as a candidate bottleneck rather than trying to reconstruct the full cross-system view.
What's the difference between process mining and task mining?
Process mining works on structured event logs from systems of record — timestamped records of which case hit which system state, when. It's best for workflows that run primarily inside enterprise software. Task mining works on screen recordings — it captures what users actually do on their desktops, including operations that happen outside of system logs (copy-paste from email, Excel manipulation, browser forms). Process mining is faster to set up, cheaper to run, and produces cleaner data. Task mining is necessary when the workflow lives partly or wholly in desktop applications that don't emit structured logs. Most organizations should start with process mining and add task mining only for specific knowledge worker workflows where system log coverage is low.
How does this connect to picking our first three AI workflows?
Process mining is the upstream data layer for the workflow selection framework covered in the companion article on picking your first three AI workflows. That framework (risk × value × signal quality) tells you how to structure your first three pilots once you have candidates. Process mining tells you which candidates to put into that framework in the first place. The two approaches are designed to be used in sequence: mine first to generate a ranked candidate list, then apply the selection framework to decide which three candidates to commit to as pilots. Using the selection framework without mined data means you're scoring workshop opinions rather than measured process data — which is where most first AI programs go wrong.
Process Mining for AI Discovery Checklist
Identified two source systems with the highest operational process volume and accessible event logs
Secured data access permissions before tool procurement (InfoSec approval, API credentials)
Selected a mining tool appropriate to your stage: Disco/Apromore for discovery sprint, Celonis/IBM for enterprise deployment
Extracted and validated event logs: case ID, activity/status, and timestamp fields all present and complete
Rendered variant analysis and identified top 20 workflows by accumulated wait time
Scored each workflow on Volume × Variance × Structure (1–5 per axis, multiplied)
Identified workflows above score 40 as AI candidates and above score 70 as priority pilots
Ran pre-mortem on top 5 candidates: data access, ownership, compliance, integration cost, success metric
Enrolled a C-suite sponsor before committing to the mining program (not just informed — actively co-owning)
Confirmed legal and compliance sign-off for any regulatory-adjacent candidate workflows
Committed to top 3 pilots with explicit funding, pilot owners (not just sponsors), and 60-day go/no-go criteria
Scheduled the next 90-day mining cycle to keep the candidate pipeline fresh as processes evolve
The question your organization keeps asking — "which workflows should we automate with AI?" — has a data answer. It's sitting in your SAP tables, your ServiceNow event log, your Salesforce activity history. The workflows that are genuinely worth building are the ones with high volume, high variance, and unstructured inputs: the cases where an LLM's ability to read context and handle exceptions outperforms any rule engine you could build. Process mining gives you the ranked list. The Volume × Variance × Structure score gives you the tool-selection logic. The pre-mortem tells you which ones you can actually ship. None of this requires a workshop.
Stop building AI use cases out of opinion. Your event logs already know which workflows are painful, repetitive, and ripe for automation. Mine them, score them, run the pre-mortem, then pick. The team that builds the wrong thing confidently is still building the wrong thing — and the logs have been trying to tell you that from the start.
- [1]Celonis Named a Leader in Process Mining Market Assessment — Celonis Blog(celonis.com)↩
- [2]Celonis process intelligence turns enterprise AI into ROI — SiliconANGLE, February 2026(siliconangle.com)↩
- [3]6 trends shaping process mining in 2026 — Process Excellence Network(processexcellencenetwork.com)↩
- [4]Celonis Process Mining: Products, Features & Competitors — AIMultiple Research(research.aimultiple.com)↩
- [5]What is Process Mining? — IBM Think(ibm.com)↩
- [6]Top Celonis Process Mining Alternatives: Signavio, Apromore, ARIS, and ABBYY Timeline — mindzie(mindzie.com)↩
- [7]Process Mining Software Market Size & Share Analysis — Mordor Intelligence(mordorintelligence.com)↩
- [8]Process Mining AI Explained: Benefits, Use Cases, and ROI — RTS Labs(rtslabs.com)↩