
Continuing our series on IT Service Management (ITSM) problem-solving techniques, this deep dive examines how the Pareto Analysis (80/20 Principle) can be applied within ITIL-based Problem Management. We will explore the theoretical foundations of the Pareto Principle and demonstrate how focusing on the “vital few” causes helps IT organizations efficiently identify root causes, prioritize recurring incidents, and implement cost-effective resolutions. The article also compares Pareto Analysis with other root cause methods (like 5 Whys, Ishikawa diagrams, and FMEA), discusses its use across related ITSM functions (Incident Management, Major Incident Reviews, SLA breach analysis, Continual Service Improvement), and addresses practical considerations such as data quality, tooling integration, and governance best practices. In doing so, we aim to equip ITSM professionals, infrastructure leaders, and service managers with a comprehensive understanding of how the 80/20 rule can drive continuous improvement in service reliability and performance.
Vilfredo Pareto, an Italian economist, first observed in 1906 that roughly 80% of the land and wealth in Italy was owned by 20% of the population (Giva, Inc., n.d.). This observation was later generalized into what is now known as the Pareto Principle or 80/20 Rule: in many situations, about 80% of outcomes stem from 20% of causes (Giva, Inc., n.d.) (Investopedia, n.d.). In other words, a small subset of causes (the “vital few”) often accounts for the majority of the impact, while the remaining majority of causes (the “trivial many”) each contribute only a minor amount. Renowned quality management pioneer Joseph Juran popularized this concept in business, referring to it as the “law of the vital few” and advising managers to focus on those key factors that would yield the most significant improvements (Giva, Inc., n.d.).
It’s important to note that the 80/20 split is not a rigid law but rather an empirical pattern. The exact ratio may vary (70/30, 90/10, etc.), but the core idea is an uneven distribution between cause and effect (BetterExplained, n.d.). Whether in economics, quality defects, or IT incidents, experience shows that outcomes are often not evenly distributed among all contributors (BetterExplained, n.d.). For instance, 20% of software bugs might cause 80% of crashes, or 20% of customers might generate 80% of support tickets. The Pareto distribution curve is typically visualized using a Pareto chart: a bar graph where causes are sorted in descending order of frequency (or impact) and a cumulative percentage line highlights how quickly the top causes accumulate to a large share of the effect (ManageEngine, n.d.; New Relic, n.d.).
From a theoretical standpoint, the Pareto Principle encourages a strategic focus on the areas that yield the highest payoff. Rather than spreading efforts equally across all issues, it suggests identifying which “critical few” factors are driving most of the problem and addressing those first for maximum benefit (Mobile2B, n.d.; University of Edinburgh, 2020). This principle has proven valuable in countless fields – from improving industrial quality (eliminating the most frequent defect types) to managing personal time (spending time on the 20% of tasks that create 80% of value) (Float, n.d.). In the following sections, we will see how this principle translates to IT Service Management and problem-solving in an ITIL context.
IT organizations have long embraced the 80/20 rule as a lens to improve operational efficiency and service quality. In IT Service Management, the Pareto Principle often manifests in observations such as “20% of services or components cause 80% of IT problems” or “20% of incident types account for 80% of the ticket volume” (Zenkins, n.d.). By identifying these dominant factors, IT leaders can better allocate resources and time to what matters most. As one industry publication notes, whether you are working with ITIL, Agile, or other methodologies, the Pareto Principle can be a powerful tool for success (Giva, Inc., n.d.). It helps answer the crucial question: where should we focus our improvement efforts to get the biggest return?
ITIL Problem Management, in particular, benefits greatly from Pareto Analysis. ITIL (Information Technology Infrastructure Library) provides a framework for managing IT services, and Problem Management is the process responsible for investigating and eliminating the root causes of incidents. The goal is to reduce incident recurrence and minimize the impact of issues by addressing underlying problems rather than just treating symptoms. Within this context, Pareto Analysis provides a data-driven way to prioritize which problems to tackle first (ManageEngine, n.d.). Rather than arbitrarily picking problems or relying solely on intuition, IT service managers can analyze incident data to discover which problems or causes are generating the bulk of incidents or downtime.
For example, an ITSM team might find that out of hundreds of incidents last quarter, a small number of recurring issues (such as VPN connectivity failures or a particular software error) constituted the majority of disruptions. By applying Pareto Analysis, “the team can improve productivity, reduce costs, and maximize ROI” by focusing on those key areas (Giva, Inc., n.d.). In the words of the Giva ITSM team, 20% of projects or services could fix 80% of IT problems if identified correctly (Zenkins, n.d.). Efficiency in ITSM is often about working smarter, not just harder – Pareto analysis helps pinpoint where improvements will yield the most significant outcomes (Mobile2B, n.d.).
Crucially, the 80/20 rule is not a rigid prescription but a guiding principle. ITSM leaders are advised to use it thoughtfully. As the Giva authors caution, it’s not a “law” but a useful theory that, when applied properly, “can work wonders for operational productivity” (Giva, Inc., n.d.). In practice, the exact percentages will vary, and sometimes an outsized effort might be required for a critical issue that isn’t very common (more on that in the limitations section). Nevertheless, the underlying lesson remains: identifying the dominant causes of pain in IT services allows teams to focus their limited resources for maximum impact (Mobile2B, n.d.; University of Edinburgh, 2020).
Next, we will zoom into ITIL Problem Management and see how Pareto Analysis is integrated into problem-solving practices, helping to identify root causes and prioritize solutions in a structured way.
Problem Management in ITIL aims to find permanent solutions to the root causes of incidents and thus prevent recurrence. Within this process, after incidents are analyzed and problems (underlying causes) are identified, teams must decide which problems to address first – especially when faced with many issues and limited resources. This is where Pareto Analysis becomes invaluable as a prioritization technique. In fact, many ITIL practitioners explicitly include Pareto Analysis as one of the recommended problem management techniques, alongside methods like root cause analysis (RCA) and brainstorming (Purple Griffon, n.d.-a). By “using the Pareto principle (80/20 rule) to identify the most significant problems and prioritize them for resolution”, problem managers can ensure they tackle the issues that will yield the largest improvement in service stability (Purple Griffon, n.d.-a).
How Pareto Analysis Supports Problem Management: First, it helps identify recurring incident patterns. Problem Management often begins by sifting through incident records to detect problems (a problem is typically defined as the unknown root cause behind one or more incidents). Through Pareto Analysis of incident data, the ITIL practice can highlight which categories or symptoms are most frequent. For instance, if 40% of all helpdesk tickets are related to email service outages, and the next 20% are related to VPN issues, those emerge as prime candidates for deeper problem investigation. ITIL’s guidance emphasizes focusing on incidents “that have significant impact or frequency” when identifying problems (Tempo, n.d.). Pareto charts make such decisions fact-based: the data might show, say, that just two categories of incidents account for ~60-70% of all occurrences, indicating a likely payoff if those are resolved (Purple Griffon, n.d.-b).
Secondly, Pareto Analysis aids in root cause identification and verification. Once the major problem areas are identified, problem analysts use root cause analysis techniques (5 Whys, Ishikawa diagrams, etc.) to find the underlying reasons. Here, Pareto doesn’t replace those analytical techniques but complements them. A ManageEngine ITIL guide succinctly states: “This analysis complements the Ishikawa and Kepner-Tregoe methods by providing a way to prioritize the category of problems, while the other methods analyze the root cause.” (ManageEngine, n.d.). In other words, Pareto tells us what to investigate first, while methods like 5 Whys and Fishbone tell us why it’s happening (ComplianceQuest, n.d.). By starting with the most prevalent problem category, IT teams can spend their limited diagnostic time on areas likely to prevent the most incidents.
Thirdly, Pareto Analysis supports cost-effective resolution and continual improvement. By addressing the top few root causes, IT organizations can “significantly reduce service disruption” across the board (ManageEngine, n.d.). For example, eliminating a single recurring database glitch that caused 200 incidents per year has a clear ROI compared to fixing a very rare bug that caused 2 incidents. One ITIL case study found that implementing a self-service password reset (to tackle one of the most frequent service desk tickets) and educating users on common mistakes led to a substantial drop in total incident volume (Purple Griffon, n.d.-b). This illustrates the classic 80/20 payoff: by focusing on the ‘vital few’ issues, the IT team achieved outsized improvements in service quality and workload reduction (Purple Griffon, n.d.-b). Such outcomes feed into ITIL’s Continual Improvement practice, where Pareto charts can measure improvement (e.g., seeing the share of incidents from top issues shrink after fixes are applied).
To use Pareto analysis within problem management, a systematic approach is followed:
Gather incident and problem data over a period (e.g. last month or quarter). Incidents should be categorized by relevant attributes – for example, by cause, by affected service, by configuration item, or by symptom. The accuracy of categorization here is critical (more on data quality later). ITIL tools often allow tagging incidents with categories or cause codes once known errors are identified. “Collect problem ticket data from your service desk tool” and organize it into countable classifications, advises ManageEngine (ManageEngine, n.d.).
Create a Pareto table that lists each category of interest and the number of incidents (or problems) in that category (ManageEngine, n.d.). Then sort categories from highest to lowest count. Calculate the cumulative percentage contribution of each category to the total incident count (ManageEngine, n.d.). Often, this is where one discovers, for example, that the top 3 categories constitute, say, 75% of all incidents.
Visualize the results as a Pareto chart – a bar chart sorted by frequency with a cumulative percentage line. This chart makes it easy to see the break-point where the cumulative line flattens, indicating diminishing returns. Figure 1 below provides an example of such a Pareto analysis for IT incident causes.

Figure 1. Pareto Analysis of Incident Causes (Example)
Example Pareto chart of IT incidents by cause: In this scenario, “Password Resets” and “User Errors” are the top two incident causes, together accounting for roughly 64% of all tickets (as indicated by the orange cumulative line reaching ~64% at the second category). Such insights help ITIL Problem Management focus on developing solutions (like self-service password reset or user training) that could eliminate a large volume of repetitive incidents (Purple Griffon, n.d.-b).
Using the chart or cumulative data, determine which causes make up roughly ~80% of the impact. These top causes are the “vital few” that should receive priority. In our example above, one can see a clear drop after the second or third bar, suggesting those first two categories dominate the incident volume. In practice, “the Pareto chart reveals that [the top causes] together account for approximately 70% of all tickets” (Purple Griffon, n.d.-b) – a strong indication that addressing them would dramatically reduce workload.
For each of the top categories, initiate Problem Management investigations if not already done. Use root cause analysis techniques to find the underlying root cause(s) of those frequent incidents. For instance, if “Connectivity issues” is a top category, the problem investigation might find that a specific network switch firmware bug is the root cause of repeated outages.
Develop permanent fixes or improvement actions for those root causes. ITIL Problem Management works closely with Change Management here, since implementing a fix may require a change in the infrastructure. The Pareto approach ensures Change initiatives are justified by impact. (For example, implementing redundancy for a frequently failing server will have clear incident reduction benefits as shown by the data.)
After deploying fixes, measure incident trends again. The expectation is that incidents related to the resolved problems should drop significantly, thereby validating the ROI of focusing on those areas. If successful, the next “vital few” issues may bubble up as new priorities.
It’s worth noting that ITIL Problem Management distinguishes between reactive and proactive problem management. Reactive Problem Management addresses problems that have caused incidents; proactive identifies problems before incidents occur (by analyzing trends, events, etc.). Pareto Analysis is powerful in both modes. In reactive mode, it uses historical incident data to drive root cause elimination for recurring issues. In proactive mode, an IT team might apply Pareto analysis to event logs or minor incident trends to catch a growing issue early. For example, analyzing non-critical error tickets might reveal a trend to fix before it becomes a major incident.
Finally, Pareto Analysis is often used in ITIL Problem Management during Major Incident Reviews or post-mortems. After a major outage, problem managers will catalog all contributing factors and often find that a combination of a few factors caused most of the impact. While a major incident is a single event (not a frequency distribution of many events), Pareto thinking can still apply in lessons learned: address the biggest contributors to downtime first in remediation plans. Additionally, if multiple major incidents are reviewed together, Pareto charts can categorize their causes to see if, for instance, 20% of infrastructure components are responsible for 80% of major incident minutes – valuable insight for reliability engineering.
In summary, Pareto Analysis is embedded in ITIL Problem Management as a means of targeting the most impactful problems for root cause analysis and resolution (ManageEngine, n.d.). By doing so, IT organizations can maximize the reduction in incidents and downtime with the effort available. The next section will illustrate this with a concrete example case, and then we’ll broaden the view to other ITSM processes.
To make the discussion more concrete, let’s walk through a realistic ITSM case example where Pareto Analysis is used to identify and resolve the dominant causes of recurring incidents. Suppose an enterprise IT Service Desk is experiencing a high volume of support tickets, leading to backlogs and missed Service Level Agreement (SLA) targets. The Problem Manager decides to perform a Pareto Analysis on the incident data from the past month to pinpoint what’s driving most of the tickets.
Over one month, the team collected and categorized all incoming incidents by their primary cause. The categories and ticket counts came out as follows:
In total, 1,400 incidents were logged in the month. Already, the raw numbers hint at an imbalance: two categories (password resets and user errors) together account for 700 incidents, which is 50% of all tickets.
The team plots these on a Pareto chart (as illustrated earlier). The bars (in descending order) show the incident counts per category, and the cumulative percentage line climbs steeply for the first few categories. The chart clearly indicates that password resets are the single largest category, and when combined with user errors, they constitute roughly 70% of all tickets (Purple Griffon, n.d.-b). This aligns with the Pareto Principle expectation that a minority of causes produce the majority of effects. In fact, by the time we include the third category (network issues), the cumulative line passes roughly ~88% – meaning just three causes are nearly 90% of the workload, while the remaining three causes (software bugs, hardware, performance) make up only ~10-12% collectively.
From this analysis, the Problem Management team identifies “Password resets” and “User errors” as the top two problem areas to tackle immediately. These are the vital few causing the most frequent pain. Network issues, while also significant, are third in line and may be addressed in the next phase.
The team then conducts root cause analysis for these top issues:
For both categories, solutions are brainstormed and implemented:
After implementing these actions, the team closely monitors the ticket volumes over the next few months. The results are impressive: password reset tickets drop dramatically as users adopt the self-service tool (only occasional calls come in for exceptions), and user-error related tickets also see a steady decline due to the combination of better documentation and perhaps smarter UI changes in some apps. The service desk reports a tangible reduction in daily ticket load. In fact, if password resets and user confusion are largely mitigated, the overall incident volume could drop by nearly 50%, corresponding to our identified ~700 tickets/month – a massive efficiency gain. This also means service desk agents can respond faster to other issues and spend more time on complex incidents, improving SLA compliance and user experience.
For completeness, the team proceeds to address the next items (network issues, software bugs) in subsequent Problem Management cycles. But thanks to Pareto prioritization, they addressed the highest pay-off problems first, yielding immediate relief. A senior IT manager reviewing this outcome might note that Pareto Analysis “helped the ITSM team prioritise efforts to maximise impact on service quality and efficiency” (Purple Griffon, n.d.-b) – exactly as intended. It showcases how focusing on the 20% of causes (two categories here were ~33% of categories, which gave ~70% of effect) can dramatically improve IT operations.
In this case study, Pareto Analysis provided clarity on which issues were causing the majority of incidents, and thus where improvements would be most cost-effective. The solutions (self-service tool, user training) not only reduced ticket counts but also had secondary benefits: freeing up IT staff, improving user satisfaction, and potentially saving costs (fewer calls means lower support cost). This kind of data-driven decision-making is at the heart of ITIL’s continual improvement ethos. It also illustrates the interplay between Problem Management and other practices: the solutions involved Change Management (implementing a new tool), Knowledge Management (improving user guides), and Incident Management (monitoring metrics like incident volume and resolution time improvements).
With a solid understanding of how Pareto Analysis can zero-in on major pain points in Problem Management, we now broaden our scope to related ITSM functions where the 80/20 principle can be applied in a similar fashion.
Beyond Problem Management, the Pareto approach is useful in several other ITIL and ITSM processes for analysis and improvement. Let’s explore a few key areas:
Incident Management focuses on restoring normal service operation as quickly as possible when an incident occurs. While its primary mandate is not root cause removal (that’s Problem Management), Incident Management can still leverage Pareto Analysis to identify high-frequency incident types and feed the Problem Management process. Service desk tools often have reporting modules that show, for example, the top 10 incident categories or trending issues. Using Pareto charts on incident data helps Incident Managers and service desk leads recognize when a small number of issues are causing a large chunk of workload or user impact (Tempo, n.d.). For instance, the service desk may notice that “password resets” or “VPN not connecting” are repeatedly in the top incidents list – a signal to initiate a Problem ticket for those. Many ITSM teams set up dashboard widgets for top incident causes; this is effectively real-time Pareto analysis to drive prompt action.
Furthermore, Incident Managers during periodic reviews can use 80/20 thinking to improve efficiency: If 20% of support agents handle 80% of the tickets, perhaps their methods or knowledge should be replicated to others (or workload balanced). If 20% of knowledge base articles are used in 80% of ticket resolutions, ensure those key articles are easy to find and kept up to date. All of these exemplify focusing on high-impact elements in the incident support process.
After a major incident or outage, ITIL recommends holding a Post-Incident Review (PIR) or major incident review. The goal is to understand what went wrong, what contributed to the impact, and how to prevent similar incidents. In these reviews, Pareto Analysis can be used to analyze contributing factors and downtime. For example, consider a prolonged system outage that had multiple causes (hardware failure, delayed failover, human error in response, etc.). By analyzing the timeline and impact of each contributing factor, one might find that a single factor (say, a misconfigured failover that delayed recovery) accounted for 80% of the downtime. That factor should be the top priority for remediation.
If reviewing multiple major incidents over a period, Pareto charts can categorize causes across incidents. A real-world example: suppose over a year, the IT org had 5 major incidents: two caused by storage failures, one by network, one by a software bug, and one by human error. The Pareto chart would show storage issues as the top contributor (40% of major incidents in count, perhaps more in impact). This insight would push the team to strengthen storage reliability (perhaps via better redundancy or vendor changes) as a first step in the major incident reduction program.
Anecdotally, many organizations find that a few systemic weaknesses cause repeated high-severity incidents. Pareto Analysis during PIRs highlights these systemic issues so that Problem Management can jump on them. It aligns with an ITIL guiding principle to “focus on value” – in this case, preventing the major incidents that cause the most business damage by addressing top causes first (University of Edinburgh, 2020).
Service Level Management is about ensuring that IT services meet agreed-upon performance targets (SLAs). When breaches occur (e.g., uptime below target, response times not met, resolution times exceeded), analyzing the pattern of breaches can benefit from Pareto techniques. IT service managers often perform Pareto Analysis on SLA breaches to see which services or which failure modes cause the majority of SLA violations. For example, if an organization has 50 services with SLAs but finds that “20% of the services are responsible for 80% of SLA breaches,” that tells management where to focus improvement projects (perhaps those services need upgrades, more capacity, or better monitoring) (Luxembourg Institute of Science and Technology, n.d.).
Similarly, within a single service SLA, one could analyze causes of downtime. Perhaps out of all downtime minutes in a quarter, one cause (a memory leak crash) accounts for most of it. That cause should be the focus of a Problem Management fix to improve SLA compliance. As KnowledgeHut’s ITIL guide notes, “By understanding the underlying reasons for SLA breaches, organizations can implement targeted improvements to meet or exceed agreed-upon service levels.” (KnowledgeHut, n.d.). This sentiment underscores using data to drive improvement where it matters most for customer experience and contractual obligations.
Continual Improvement is a practice in ITIL that seeks ongoing enhancement of services and processes. Pareto Analysis is practically synonymous with CSI’s ethos of focusing on high-impact improvements. In fact, a common debate in CSI is whether to pursue many small “marginal gains” improvements or a few big impactful improvements. The answer, as one ITIL blog humorously put it, is “Yes… it depends.” (University of Edinburgh, 2020). The Pareto approach advocates starting with those big ticket items that provide significant ROI – which aligns with the ITIL guiding principle “Focus on Value” (University of Edinburgh, 2020). The idea is, if you have limited resources for improvement initiatives, tackle the changes that will deliver the greatest value first (the Pareto vital few), then move on to secondary ones (University of Edinburgh, 2020).
For example, in CSI one might have a register of improvement opportunities: process A could be slightly optimized (minor gain), or process B has a known bottleneck that if removed would save a lot of time (major gain). Pareto thinking would prioritize process B’s improvement because it yields more value. As the University of Edinburgh’s ITIL team discussed, one should “work on improvements that provide the best return on investment (ROI) – the vital few in the first instance. Once these high consequence improvements have been made, then address the next tier… eventually reaching marginal gains.” (University of Edinburgh, 2020). This staged approach ensures quick wins on major pain points.
CSI also uses lots of metrics. Pareto charts can identify, for instance, which processes have the most deviations, which service metrics are most often missed, or which departments generate the most complaints. By regularly reviewing such data, a CSI manager can direct the continual improvement cycle to the areas that will most improve overall quality or customer satisfaction. It’s a way to avoid the trap of spending effort on improvements that don’t materially move the needle.
Change Management (now often called Change Enablement in ITIL4) could even apply Pareto thinking to changes and incidents. For instance, if 20% of changes cause 80% of the disruption or failed change incidents, identify those change types or teams and tighten the process for them. Or, identify the few infrastructure components where most changes cluster (maybe a core switch, etc.) and ensure they have extra testing. This crosses into risk management: focus on the riskiest elements.
In analyzing failures, sometimes a Pareto chart of assets (configuration items) by incident count is revealing. It might show, for example, that one particular server or one particular software module is linked to a disproportionate number of incidents (the “bad apple” scenario). This could justify replacing that component or rearchitecting it. Similarly, in asset management, Pareto is used in inventory control (e.g., the classic ABC analysis in inventory is essentially Pareto-based, focusing on the 20% of items that constitute 80% of value).
In summary, Pareto Analysis has broad applicability across ITSM for any situation where understanding the imbalance of contributions can guide decisions. It is particularly powerful in processes that involve analysis of failures, improvements, or performance data. Incident and problem trends, SLA metrics, customer feedback, and change success rates can all be put through an 80/20 filter to inform strategy.
However, while Pareto charts tell us where to look or what to prioritize, they do not tell us why something is happening. For that, ITSM professionals employ various root cause and analysis techniques. In the next section, we will compare Pareto Analysis with some of those other popular techniques – highlighting how they differ and complement each other in the toolkit of problem-solving.
In ITSM problem-solving, Pareto Analysis often works in conjunction with other Root Cause Analysis (RCA) and problem prioritization methods. Each technique has a distinct purpose and strength. Here, we compare Pareto Analysis with three commonly referenced techniques – the 5 Whys, Ishikawa (Fishbone) Diagrams, and Failure Mode and Effects Analysis (FMEA) – to understand their roles, differences, and how they can complement each other.
As discussed, Pareto is all about using historical data to identify the “vital few” causes contributing most to a problem (typically measured by frequency or impact) (KnowledgeHut, n.d.). Its primary use is prioritization – it ranks issues so you focus on the most significant ones first (ComplianceQuest, n.d.; Mobile2B, n.d.). What it doesn’t inherently do is dig into why those issues occur. Pareto analysis yields a list of biggest contributors, but once you pick a specific issue from that list, you need other RCA methods to find its root cause (Purple Griffon, n.d.-b). Think of Pareto as pointing you to “which hill has the most gold” in a gold mining analogy; it tells you where to dig, but not how deep or in which direction – that’s for other tools.
The 5 Whys technique involves repeatedly asking “Why?” (around five times, or as many as needed) to drill down from an incident’s symptom to its root cause (ManageEngine, n.d.; Zenkins, n.d.).
For example, an incident might be “server outage”
– Why? Because of a power failure
– Why? Because the UPS failed
– Why? Battery was past end-of-life
– Why? Maintenance procedures were not followed, and so on, until you reach the fundamental cause.
5 Whys is a straightforward, qualitative technique best suited for single-problem analysis (Zenkins, n.d.). It shines in scenarios where a specific incident or problem needs cause-and-effect interrogation. However, 5 Whys is not great at handling complex or multiple concurrent causes, and it doesn’t quantify anything. As such, 5 Whys complements Pareto by being a tool you apply after Pareto flags a problem area. You might use Pareto to identify that “UPS failures” are a top cause of incidents, then use 5 Whys on a representative UPS failure incident to find the root cause (e.g., maintenance process failure).
Another point is 5 Whys is reactive and incident-specific, whereas Pareto can analyze aggregated data. If you have many incidents, doing 5 Whys on each would be tedious; Pareto summarizes which categories to investigate. Notably, “the 5 Whys method is especially useful when there is no evident root cause, while Pareto helps to grade known causes and prioritize responses” (ComplianceQuest, n.d.). In practice, an organization might document a quick 5 Why analysis for each major incident, but then use Pareto to see which root causes are most frequent across incidents.
The Fishbone diagram is a visual tool to identify many possible causes for a problem and categorize them (e.g., into categories like People, Process, Technology, Environment) (ManageEngine, n.d.). It looks like a fish skeleton: the “head” is the problem, main “bones” are categories, and smaller “bones” are sub-causes. Fishbone analysis is excellent for comprehensive brainstorming of potential causes – especially when a problem is complex or multifaceted (ManageEngine, n.d.; InvGate, n.d.). In ITIL Problem Management, fishbones are often drawn during problem investigation workshops to make sure all possible angles are considered (network, software, human error, etc.) (ManageEngine, n.d.).
Comparatively, Fishbone and Pareto serve different stages of analysis. Fishbone is exploratory, used to find root cause hypotheses in a structured way, often before data is fully known (InvGate, n.d.). Pareto is analytical, used to quantitatively confirm which causes are most prevalent or significant after data collection (ComplianceQuest, n.d.). In fact, fishbone and Pareto are frequently used together: “Fishbone and 5 Whys allow teams to identify multiple causes and drill down, while Pareto helps grade and prioritize each cause’s importance” (ComplianceQuest, n.d.). One might create a fishbone diagram to map out all candidate causes of “service downtime,” then as data comes in (incident counts, etc.), apply Pareto to see which branches of the fishbone are causing most incidents.
Another difference: Fishbone is qualitative and visual, great for team collaboration and seeing cause-effect relationships (ComplianceQuest, n.d.). Pareto is quantitative and tabular/graphical, great for seeing hard numbers and making decisions on resource allocation (Mobile2B, n.d.). Fishbone can occasionally overwhelm or become very complex if too many causes are thrown in (one noted risk is “digression and confusion due to too many ideas” on a fishbone) (ComplianceQuest, n.d.). Pareto, by contrast, simplifies the view to a ranked list. Thus, they balance each other: fishbone ensures no stone is unturned, Pareto ensures the most important stones are addressed first.
FMEA is a systematic technique originally from engineering to identify all the ways something could fail (Failure Modes), and analyze the effects of those failures on the system (Mobile2B, n.d.). Each failure mode is scored on three factors: Severity (impact of the failure), Occurrence (likelihood of it happening), and Detection (ability to detect/prevent it). The product of these (Risk Priority Number, RPN) helps prioritize which failure modes to address first (Mobile2B, n.d.). FMEA is highly quantitative and proactive – it’s done before failures occur (or to prevent further failures) and is very detailed and resource-intensive.
Comparing FMEA to Pareto:
In practice, Pareto and FMEA can complement each other in IT risk management. Suppose an IT team used Pareto on incidents and found the top issues; then they apply FMEA on a critical service to identify a high-severity failure mode that hasn’t happened yet but could be catastrophic. They would be wise to address both: fix the frequent issues (to reduce annoyance and cost) via Pareto prioritization and proactively mitigate the critical potential issue via FMEA insights. This aligns with ITIL’s balance of reactive vs. proactive problem management – Pareto often drives reactive improvements (based on past incident data), while FMEA is a proactive tool (avoiding future incidents) (InvGate, n.d.).
A concise comparison:
These tools are not mutually exclusive. In fact, ITIL literature and experts encourage using them in combination: “The five whys technique complements many other problem-solving techniques like the Ishikawa method, Pareto analysis, and the Kepner-Tregoe method.” (ManageEngine, n.d.). For example, one might identify a hotspot via Pareto, then use a Fishbone with a team to identify causes, then apply 5 Whys on a specific branch of the fishbone to drill down to root cause, and if designing a solution, use FMEA to ensure the solution doesn’t introduce new failure modes.
A real-world sequence could be:
The solution becomes “implement standard deployment checklist and better documentation”. Before rolling that out, the team could apply FMEA to the new process to anticipate any new failure modes (e.g., if the checklist is too cumbersome, people might skip steps – so they might simplify it). In this way, each technique plays a role: Pareto prioritizes the problem to solve, Fishbone and 5 Whys uncover the solution, and FMEA polishes the solution for robustness.
While not asked in the question, it’s worth noting ITIL Problem Management also mentions structured techniques like Kepner-Tregoe (a step-by-step analytical approach) and Fault Tree Analysis (visual logic diagram of causes) (InvGate, n.d.). Pareto differs from those too – KT is about systematically evaluating information to pinpoint causes (good for single problems), and Fault Tree is a deductive model of how multiple lower-level failures can combine to cause a higher-level failure (used a lot in safety engineering). Those are beyond our scope here, but they too can be interwoven (e.g., Pareto could identify which problem to run a KT analysis on first).
To wrap up this section: Pareto Analysis is not a root cause finding tool in itself; it is a root cause prioritization tool (Purple Griffon, n.d.-b). It excels when you have multiple issues and need to decide where to focus. Techniques like 5 Whys and Ishikawa are about digging into a specific issue to find the cause. FMEA is about preventing issues by tackling high-risk potential failures. Knowing which to use when is crucial for effective problem management. As one quality management blog noted: “If a problem has many root causes, Fishbone and 5 Whys can help to unearth them (and organize them), while Pareto helps to grade the known causes and prioritize the response to each.” (ComplianceQuest, n.d.). And importantly, the three can be used together – e.g., fishbone to visualize cause categories, 5 Whys to drill into each, and Pareto to focus on the most significant categories (ComplianceQuest, n.d.).
Finally, one must consider the organization’s context: for simple issues, 5 Whys alone may suffice; for large volumes of data, Pareto is invaluable; for critical systems, FMEA is worth the effort. Using the right tool for the right problem – and often a combination – is a best practice of mature IT problem management.
While Pareto Analysis is a powerful and popular technique, it is not without limitations. Understanding these limitations is vital to avoid misapplication of the 80/20 rule in IT service environments (Purple Griffon, n.d.-b). Here we outline common pitfalls and prerequisites to use Pareto Analysis effectively in ITIL-based problem management:
Perhaps the most important caution is that Pareto charts typically rank causes by frequency (count of incidents), which may not align with severity or business impact (Purple Griffon, n.d.-b). In IT, not all incidents are equal – one rare outage affecting a payroll system during year-end might be far more damaging than ten frequent minor glitches in a training application. Pareto Analysis can inadvertently downplay high-impact but infrequent issues. For instance, if a critical security breach happened once, it might appear trivial in a yearly incident Pareto chart compared to dozens of password reset incidents; yet preventing another breach could be far more important. To mitigate this, IT teams should augment Pareto with impact analysis – e.g., weighting incidents by their severity or cost. Alternatively, perform separate Pareto analyses for frequency and for total downtime caused, and ensure that you consider both perspectives. In short, don’t let volume metrics alone drive decisions; consider impact and risk. (This is where frameworks like FMEA or risk matrices complement Pareto, as discussed earlier.)
Pareto is only as good as the underlying data quality. In ITSM, this means incidents must be properly categorized, and relevant metadata (like cause codes or resolution codes) should be consistently recorded. A common challenge is unreliable or inconsistent incident categorization, which can lead to misleading Pareto charts (Purple Griffon, n.d.-b). For example, if support technicians categorize half of the tickets as “Miscellaneous” or if the categories are too broad/narrow, the analysis might not reveal true root causes. Prerequisite: Implement clear categorization schemes (such as standardized incident classification in the ITSM tool) and training for staff to use them correctly. Some organizations invest in data cleanup or use text analytics to retroactively categorize tickets more accurately before Pareto analysis. If data is scant or unreliable, Pareto results may be less effective or even “challenging to implement” (Purple Griffon, n.d.-b). Thus, ensure you have a decent sample size and trustworthy data before drawing conclusions.
As emphasized, Pareto Analysis does not tell you the root cause of the problems – it identifies symptoms or problem categories that are most frequent (Purple Griffon, n.d.-b). One must be careful not to stop at “the top category is X, so let’s just address X superficially.” Within that category, there could be multiple underlying causes. For example, “network issues” might be a top category, but the actual root causes could be diverse (router bugs, Wi-Fi interference, ISP outages, etc.). Misapplication would be trying to solve a category without proper analysis – e.g., throwing hardware at “network issues” when maybe it was a software bug. Always follow up Pareto identification with deeper RCA on those top categories. Pareto tells you where to focus, but you still need to determine what solution will eliminate that cause (ManageEngine, n.d.).
Real-world IT problems often have multiple interrelated causes. Pareto Analysis can sometimes oversimplify these by focusing on individual cause categories in isolation (Purple Griffon, n.d.-b). For instance, a major incident might have occurred due to a combination of factors, not just one cause. If you only address one without the other (because it showed up slightly higher in Pareto ranking), you might only partially fix the issue. Additionally, some issues might not be independent – e.g., “power failure” and “UPS failure” might both appear, but they are linked (one triggers the other). Pareto won’t show that relationship. It treats categories as separate and rankable, which is a simplification. ITIL practitioners should be wary of this and ensure a holistic view: use Pareto as a starting point but then analyze relationships and systemic factors among top causes (this is where tools like Fishbone, Fault Tree, or just expert judgment come in). Avoid a siloed approach to fixes that ignores system interactions.
There is a risk of confirmation bias or misinterpretation with Pareto charts. For example, how one defines categories can change the outcome. If categories are too granular, the top cause might not look so big; if categories are lumped, one might dominate. People may inadvertently define categories in a way that highlights what they think is the problem. Also, if data collectors know Pareto is being used, they might skew categorization (even unintentionally). Transparency and consistency in how data is binned are crucial. As a best practice, document the criteria for categorizing incidents and avoid changing definitions mid-stream (or account for it if you do). Also, treat Pareto results as one input, not gospel – cross-check with anecdotes from support staff or other metrics to ensure it makes sense.
IT environments change rapidly – new systems come online, patches fix some issues, others arise. A Pareto chart is a snapshot in time. The “vital few” this quarter might not be the same next quarter (Purple Griffon, n.d.-b). One limitation is that organizations may create a Pareto analysis, address the top problems, and then not update the analysis. The risk is either solving yesterday’s problems or missing newly emerging ones. Continual Service Improvement demands repeated Pareto analyses on fresh data – essentially making it a part of ongoing reporting. However, doing so is resource-intensive if done manually. This is where having automated reporting or dashboards helps (more in tooling section). In any case, be mindful that once you remove the top causes, the distribution shifts – there will be a new 80/20 among remaining issues. Iterate and adjust focus accordingly. Also, seasonality or one-off events can skew data (e.g., a one-time cyberattack causing 50 incidents in one month). It may dominate a Pareto for that period but not be relevant long-term; knowing context is important.
The flip side of Pareto’s advice to focus on the vital few is the risk of completely ignoring the remaining issues. While each of the “trivial many” may be small, collectively they might still account for 20% of problems (by definition, if 80% come from 20%, the other 80% of causes give 20% of problems). In a large environment, 20% of problems might still be a lot. Also, some of those lesser issues might be easy wins – quick fixes that improve user experience. If you only ever focus on the top 2 problems, you might have diminishing returns once those are largely solved, while lots of minor annoyances remain for users. The key is balance – prioritize major issues first, but don’t permanently ignore the rest (Purple Griffon, n.d.-b). ITIL’s continual improvement suggests that after big issues are addressed, you can progressively tackle smaller ones. As one ITIL blogger put it, after you address the vital few, “then we can progressively address the next tier... eventually, we’ll get to marginal gains” (University of Edinburgh, 2020). Also be aware that many smaller issues could share an underlying cause (which if grouped would have been a big issue). Make sure the trivial many are genuinely separate low-impact issues, not a fragmented view of a bigger issue.
Pareto charts rely on correct grouping of incidents under causes. If incidents are mis-grouped, the conclusions can be wrong. For example, suppose incidents are grouped by symptom (“server down” incidents vs “slow application” incidents) rather than true cause. The Pareto might show “server down” as a big category, but that’s a symptom category containing multiple causes (power, OS crash, network). Addressing “server down” generally isn’t actionable except by investigating each occurrence. Thus, define Pareto categories meaningfully – ideally actual root cause categories or at least service/component categories – not too generic. ITIL Problem Management distinguishes between symptoms and root causes; your Pareto should aim as close to root cause as the data allows (e.g., use known error categories or CI categories rather than vague terms). If the data only has symptoms, use Pareto as a starting point but then break down that symptom via Problem Management.
Another limitation is focusing on problems only in terms of incidents and failures. There may be opportunities for improvement that Pareto (which looks at problems) doesn’t highlight. For example, no one may be complaining about a process, but maybe it’s very inefficient. It won’t show up in an incident Pareto because it’s not causing incidents, yet improving it could save money or time. Organizations that chase only what Pareto shows could become too reactive. It’s important to also proactively look at industry best practices, user experience feedback, or innovation opportunities – not just incident data. In CSI terms, Pareto focuses on problem solving, but CSI also involves improving quality and value, which might involve things not in any “top 10 incident” list. Don’t let the data blind you to qualitative aspects. Combine Pareto’s findings with strategic goals: for example, perhaps an issue is not frequent but strategically critical to fix for future growth or compliance.
In summary, Pareto Analysis remains a highly valuable tool, but IT leaders must apply it with nuance and in combination with human insight. Being aware of its biases ensures that one uses it wisely rather than blindly. A Gartner analyst might say: use Pareto to drive data-driven decisions, but overlay it with business context and risk awareness. IT Governance processes (like Change Advisory Boards or Problem Management reviews) should ask questions like: “Are we addressing the top issues as indicated by data? Are there any high-risk issues not indicated by past data? Are we maintaining data quality to trust these analyses?” Answering these ensures Pareto is used in service of business value, not just chasing metrics.
To mitigate these pitfalls, best practices include: maintain good data hygiene, integrate impact/severity into analysis (e.g., by creating a weighted Pareto or two-dimensional priority matrix), revisit analyses regularly, and educate teams that Pareto is a guide, not an absolute. When done right, the limitations can be managed, allowing the organization to reap the efficiency benefits of Pareto analysis without falling into its traps.
Modern ITSM environments benefit from a plethora of tools and platforms that can automate and enhance analysis like Pareto. Integration of Pareto Analysis into tooling means faster insights, continuous monitoring, and even automated actions when certain thresholds are met. Here we explore how contemporary ITSM and analytics tools support Pareto analysis, and how automation and dashboards can embed the 80/20 principle into daily operations.
Many service management tools (ServiceNow, Jira Service Management, ManageEngine ServiceDesk Plus, Freshservice, etc.) provide out-of-the-box analytics and reports. Pareto charts or “Top N” reports are common features. For instance, ServiceNow’s analytics documentation highlights Pareto reports that identify the most important dimensions using descending bars and a cumulative line (ServiceNow, n.d.-b; ServiceNow, n.d.-a). ITSM administrators can often create a Pareto chart of incidents by category with a few clicks in these tools. Some platforms have drag-and-drop dashboard builders where you can add a Pareto visualization to show, say, top incident causes or top CI failures.
By having Pareto charts on live dashboards, IT managers and support leads can continuously track if the 80/20 distribution is shifting. For example, if a new problem is emerging, it will climb the Pareto chart and become visible perhaps even before a formal analysis is done. This supports proactive problem identification. Automated reporting can also schedule a Pareto analysis email every week or month to the Problem Management team.
As an example, one case study from a vendor might say: Using our ITSM dashboard, a Pareto chart revealed that 3 knowledge base topics accounted for most knowledge gap related incidents (ServiceNow, n.d.-c). With that, the team focused on updating those top knowledge articles. This underscores how integrated reporting drives action.
In addition to ITSM-specific platforms, general BI tools (like Microsoft Power BI, Tableau, Splunk, Elastic/Kibana, etc.) are often used to analyze IT operational data. These tools can query the incident database or event logs and produce advanced Pareto analyses. A benefit of BI tools is the ability to combine data sources or create custom metrics. For instance, one could produce a Pareto chart where causes are ranked by total downtime minutes or financial impact (combining incident counts with an estimated cost per incident). Microsoft’s Power BI community frequently discusses how to implement Pareto calculations on datasets (Microsoft Power BI Community, n.d.; Medium, n.d.). With a bit of DAX (Power BI’s formula language) or similar, teams can automatically flag which items constitute 80% of a volume.
Imagine if your ITSM system could automatically create or update Problem tickets when a Pareto threshold is exceeded. This is increasingly feasible. For example, a workflow could be: “If any incident category accounts for more than X% of incidents this month, and no Problem ticket exists for it, then create one and notify the Problem Manager.” This marries Pareto logic with automation. It ensures no major trend goes unnoticed. Some organizations implement scripts or use AI Ops tools to detect anomaly or trend patterns; these could effectively do Pareto in real-time and raise alerts.
In DevOps or SRE practices, teams often use analytics on monitoring data to prioritize reliability fixes. Here too the concept is similar: find the top contributors to system errors or latency. Automated analysis might pinpoint that “80% of errors in the log come from module Y” – then an automated issue could be opened for the dev team responsible for module Y.
Another angle is automated routing or escalation based on Pareto. If a certain category of incidents becomes dominant, perhaps the system could temporarily allocate more resources to it (e.g., route those tickets to a specialized team or trigger a major problem review if it passes a threshold). This moves into intelligent service management, where the platform helps enforce that the vital few problems are not stuck in the general queue.
Presenting Pareto analysis results in a clear way is key for governance and stakeholder buy-in. Many ITSM leaders include charts of top problem causes in their monthly operations review or in reports to IT governance boards. These often take the form of Pareto charts showing, for example, the top 5 recurring incidents and what is being done about them. By integrating this into dashboards that executives can view, it promotes a data-driven culture. Leadership can easily grasp from a Pareto chart that “a large chunk of our incidents come from these two issues,” which can support investment proposals to fix those issues (like funding a system upgrade or a refactoring project).
For instance, Gartner and industry research often recommend focusing on the highest-impact issues; showing the Pareto breakdown justifies that focus quantitatively. Some organizations turn these charts into continuous improvement KPIs: e.g., measure “% of incidents caused by top 5 problems” and aim to reduce that over time (meaning the top problems are being knocked out faster than new ones emerge).
Another interesting integration is with Knowledge Management. Some service platforms can do a Pareto analysis of knowledge usage or gaps – identifying which knowledge articles, if created or improved, could prevent the most tickets. For example, ServiceNow has had examples of Pareto charts to identify “candidate knowledge gaps for incidents” (ServiceNow, n.d.-c). If 80% of tickets could be solved by 20% of knowledge articles, ensure those articles are prominent and excellent; or if 80% of tickets have no KB article, focus on writing a few that cover them.
In a DevOps context, teams use incident post-mortem tools and tracking of “error budgets”. Pareto can surface as identifying the “top recurring failure types” in an application or “the 20% of code modules causing 80% of errors”. Developer analytics products or APM (Application Performance Monitoring) tools may highlight slow transactions or errors in Pareto fashion. Integrating these with work tracking (e.g., auto-create bugs for top error-causing functions) can tighten the feedback loop.
With AI, more sophisticated pattern detection is possible beyond straightforward Pareto, but the concept remains – AI might cluster incidents by similarity and then tell you which clusters are largest. That is like dynamic Pareto categorization. AI Ops platforms often identify “incident surge” or “common root cause” across alerts. They might say “these alerts all relate to one underlying cause which is 80% of your current alerts”. This achieves a similar outcome: highlighting the central issue to fix.
While tooling greatly helps, one should also beware of blindly trusting automated categorization. For instance, if an AI wrongly groups incidents, an automated Pareto result might point you in the wrong direction (garbage in, garbage out). It’s still important to have human oversight, especially in initial setup. Over time, as confidence in data grows, automation can take more lead.
Consider an ITSM team using ManageEngine ServiceDesk (Zoho). They leverage its built-in reports to generate a Pareto chart of problem classifications monthly. They embed this chart in a Confluence page or SharePoint for the ITIL Problem Review Board. Simultaneously, they have a Power BI dashboard that pulls data via API and calculates the cumulative percentages on the fly, highlighting in red any category that exceeded 80% threshold of cumulative contribution – basically visually indicating the cutoff of vital few. They also have set up an automation rule: if any single problem category exceeds, say, 50 incidents in a week, a trigger alerts the team (since that suggests a spike that might warrant a problem investigation before the monthly cycle). This integrated approach means the team is always aware of what the biggest issues are, and they can respond in near-real-time if something new starts dominating.
Another scenario is integrating Pareto with Service Level Management via tooling. For example, a monitoring tool might track SLA breaches and automatically generate a Pareto chart of causes of downtime after each outage. It could attach that to the incident ticket for review. Or a configuration analytics tool might Pareto-analyze changes that caused incidents, to feed back into Change Management policies (e.g., showing that “80% of our P1 incidents this quarter came from 2 changes – maybe those changes were not assessed well”, which suggests improving change risk assessment).
In conclusion, modern ITSM tools have embraced the Pareto principle by providing visualization, reporting, and even intelligent analytics that highlight the vital few factors. By taking advantage of these features, ITSM professionals can make Pareto analysis a continuous, almost real-time part of their management practice rather than a tedious manual exercise done occasionally. The result is a more responsive IT organization that quickly zeros in on emerging issues and constantly measures the effectiveness of problem resolution efforts.
However, technology is an enabler – organizations still need solid process and discipline to act on the insights. In the final section, we will discuss best practices in governance, data, and KPI management to ensure that Pareto analysis (and indeed any analysis) drives meaningful improvement aligned with business goals.
To maximize the benefits of Pareto Analysis in ITIL-based problem management, organizations should embed it within a framework of good governance, robust data practices, and relevant performance indicators. Here are best practices and recommendations for ITSM professionals to ensure the 80/20 technique truly delivers continuous improvement:
ITIL Problem Management (especially in larger organizations) should have clear governance structures – e.g., a Problem Manager role, a Problem Review Board, etc. It is important that ownership for acting on Pareto insights is defined. If a Pareto chart shows the top problems, who decides which ones to tackle and allocates resources? This could be part of a regular Problem Management meeting where the team reviews data and makes decisions. Ensure that management supports a Pareto-driven approach by allowing the Problem Management team to focus on those priorities. Governance should also incorporate the principle of “Focus on Value” (one of ITIL 4’s guiding principles) – meaning the team is encouraged to work on what provides the highest value (often aligned with Pareto results) (University of Edinburgh, 2020). Governance can mandate that, for example, any problem contributing to >10% of incidents must have a Problem ticket and action plan, or that funding for improvements is justified by showing it addresses a top contributor to downtime (so projects are aligned with Pareto findings).
Many organizations align ITSM with broader frameworks like COBIT, ISO/IEC 20000, or ITIL maturity models. These frameworks emphasize measurement and continual improvement. Pareto analysis can be part of meeting those frameworks’ controls. For instance, ISO/IEC 20000 (the IT service management standard) requires demonstrating improvement in service quality and reduction of incidents – using Pareto analysis results can be one way to show a systematic approach to this. Likewise, NIST’s guidance on incident management notes that lessons learned and root cause analysis help improve risk management (National Institute of Standards and Technology, 2023). Ensuring that a “post-mortem and trending” step is in your incident/problem process (and that it uses Pareto charts among other tools) can satisfy audit or compliance requirements for continuous improvement and risk mitigation.
As noted in limitations, without good data, Pareto is garbage-in-garbage-out. Best practice is to define a clear incident and problem categorization scheme. For instance, an ITIL shop might use multi-tier classification: Category, Subcategory, Symptom, Cause Code, CI, etc. For Pareto focusing on root causes, you ideally want to use the “Cause” or “Problem Type” field. Standardize these values – possibly aligning them with ITIL recommended categories or your environment’s specifics (e.g., categories like Hardware, Software, Network, Database, etc., and subcategories for specific services or components). Enforce their usage by the service desk and problem managers. Periodically review the categories: are they still suitable or do they need refinement? Train staff on the importance of proper classification, tying it to the outcome: show them that “if we categorize correctly, we can identify problems and fix them, which will make everyone’s life easier.” Often, service desk analysts will be more diligent if they see how the data is later used to drive improvements.
Additionally, aim to capture impact details for each incident (e.g., severity level, downtime minutes, user count affected). This allows weighting Pareto by impact if needed. Some organizations also integrate cost data (how much did each incident cost in terms of support time or business loss) – enabling a Pareto by cost.
It’s also a best practice to regularly audit your data. If you find a large number of incidents uncategorized or miscategorized, address that through coaching or adjusting the process (maybe make certain fields mandatory, or simplify the choices to reduce errors).
To know if Pareto-driven problem management is working, define some Key Performance Indicators. Some useful KPIs and metrics include:
In setting KPIs, ensure they are aligned with business outcomes (ITIL’s guiding principle “Keep it Simple and Practical” applies – don’t have too many) (University of Edinburgh, 2020). Many of the KPIs listed by Purple Griffon in their ITIL 4 Problem Management introduction (like number of problems, MTTD, MTTR, known errors count, resolution rate, cost of problem management) tie back to showing an effective process (Purple Griffon, n.d.-a). For Pareto specifically, one could use a KPI such as “Contribution of top 5 issues to total incidents (%); target to reduce this”, which encourages broad improvement (if one issue is gone, ideally no new one takes its place among top 5 maintaining the same high share).
Not every “top technical issue” is equally important to the business. Always interpret Pareto in light of business context and SLAs. For example, if 40% of incidents are printer jams (lots of minor incidents) and 10% are data center outages (few but huge impact), business might prioritize solving the latter even though Pareto by count says printers. In such cases, consider doing Pareto by business impact (e.g., downtime minutes, or costs) to present to decision makers. Often, IT governance boards will prioritize initiatives that reduce business risk and cost. Thus, a best practice is to convert the Pareto results into business terms for communication: e.g., “Incidents related to application X (20% of tickets) caused 500 hours of cumulative user downtime last quarter, impacting our call center. Fixing this has an estimated benefit of $Y.” This way, even if the percentages aren’t exactly 80/20, you articulate the value of focusing on the significant few problems.
Feed Pareto analysis results into the Continual Service Improvement (CSI) register (if you maintain one). A CSI register is basically a log of improvement opportunities. The ones addressing major Pareto items should be marked high priority. Conversely, if someone proposes an improvement that does not address a frequent/impactful issue, ask for justification (maybe it’s strategic or regulatory). This ensures improvement efforts align with actual service issues and not just pet projects. It’s a concrete way to implement the “Progress Iteratively with Feedback” principle – using feedback (data) to decide what to improve next (University of Edinburgh, 2020).
Promote a culture where data-driven decision-making is valued. Share successes: for example, if Pareto analysis led to fixing a big issue and subsequently incidents dropped, publicize that in IT newsletters or meetings. Show the before/after charts. This reinforces to the team that identifying root causes and solving them (rather than endlessly firefighting) yields tangible results – boosting morale and support for Problem Management. It may also encourage more people to report issues properly, knowing it leads to action. Additionally, encourage a culture of blameless problem analysis. If the top cause is “user errors,” avoid blaming users; instead focus on improving systems or training (as in our case study with education and self-service). If it’s “change failures,” avoid pointing fingers; focus on improving change process. The goal is systemic improvement, which a good ITIL culture recognizes.
Ensure that while current top issues are being handled, someone is also scanning the horizon for emerging risks (perhaps using risk registers or FMEA as mentioned). Best practice is to have both reactive Problem Management (driven by incident Pareto) and proactive Problem Management (driven by trend analysis, risk assessment). Over time, as you knock out known problems, more effort can shift to proactive identification of potential issues (which might not show in past data yet). This is how mature organizations preempt that next big incident – by not solely relying on historical Pareto, but also scenario analysis and industry knowledge.
Bake Pareto analysis into regular operational rhythm. For instance, every week or month have a report (automated as much as possible) and a meeting or at least a review by the Problem Manager. This ties back to tool integration – if it’s on a live dashboard, someone should still formally review it periodically to make decisions. Continuous monitoring helps catch shifts early. The frequency of review might depend on incident volume; high-volume environments might do it weekly, smaller ones monthly or quarterly.
Sources like NIST or ISO, or consulting frameworks, might provide benchmarks (e.g., “a high-performing service desk resolves X% of top problems per year”). Gartner often publishes research on incident trends and recommendations like focusing on recurring issues and knowledge management. Use those to validate your approach. Also, if following ITIL 4, recall that ITIL 4 emphasizes a holistic view and value streams – ensure Problem Management is connected with other practices like Change, Incident, Knowledge, and Continual Improvement (which we’ve touched on throughout) (Purple Griffon, n.d.-a). This holistic integration is itself a best practice: for example, involving the service desk and users in identifying top issues, involving change management in implementing fixes, and capturing knowledge from resolved problems for future use.
In essence, best practices for Pareto in ITSM boil down to: measure what matters, ensure the data is reliable, act on the insights, and verify the outcomes. IT governance and leadership should reinforce that process. When done well, an IT organization can confidently say: “We know where our biggest pain points are, we are addressing them in priority order, and we can demonstrate improvement through key metrics (fewer incidents, better SLAs, lower cost, happier users).” This is the ideal of ITIL Problem Management, and Pareto Analysis is a centerpiece technique to achieve it, when supported by these best practices.
Pareto Analysis, the famed 80/20 principle, proves to be a potent technique in the arsenal of ITIL-based Problem Management, enabling IT organizations to smartly channel their problem-solving efforts for maximum impact. By focusing on the “vital few” causes that contribute to the majority of incidents and service disruptions, IT teams can rapidly improve service reliability, reduce fire-fighting, and optimize the use of resources. Throughout this deep dive, we saw how Pareto Analysis supports the core goals of IT Problem Management: identifying recurring root causes, prioritizing which problems to tackle first, and implementing cost-effective resolutions that yield significant reductions in incident volumes.
We began with the theoretical foundations laid by Vilfredo Pareto and Joseph Juran, understanding that many outcomes in IT (as in life) are unevenly distributed – a small set of issues often generates most of the work (Giva, Inc., n.d.; Mobile2B, n.d.). Translating this to ITSM, we recognized that rather than treating all incidents or problems equally, focusing on the top offenders is crucial for efficiency (Giva, Inc., n.d.). In an ITIL context, this aligns perfectly with principles like Focus on Value and Progress Iteratively – deliver the biggest value improvements first, and do so in steps based on feedback (incident data being a form of feedback) (University of Edinburgh, 2020).
Through a realistic case example, we illustrated Pareto Analysis in action: by addressing two issues (password resets and user errors) that caused 70% of tickets, a service desk was able to dramatically cut down its workload and improve user satisfaction (Purple Griffon, n.d.-b). Such wins exemplify why Pareto-based problem management isn’t just about statistics – it directly correlates to better service outcomes and happier stakeholders. Real-world ITSM success stories often have this pattern: fix a handful of underlying problems and see a wave of improvement across service metrics (Purple Griffon, n.d.-b).
We also examined how Pareto Analysis extends to other ITSM processes: helping Incident Managers and Major Incident teams spot trends, guiding SLA improvements by pinpointing frequent breach causes (KnowledgeHut, n.d.), and fueling Continual Service Improvement programs with data-driven priorities (University of Edinburgh, 2020). It’s clear that the 80/20 principle is a versatile lens – whether reviewing a single outage or planning quarterly service enhancements, it informs where attention will yield the greatest return.
The comparison with techniques like 5 Whys, Fishbone, and FMEA highlighted that Pareto Analysis is not a silver bullet but a complementary component of a holistic problem-solving toolkit. Each method plays a role: Pareto often tells us what to attack, and methods like 5 Whys and Ishikawa tell us why it’s happening and how to fix it, while FMEA reminds us to consider what could happen (ComplianceQuest, n.d.; Mobile2B, n.d.). By combining these approaches, ITIL Problem Management can be both effective (solving the right problems) and thorough (finding true root causes and preventing future issues). A key takeaway is that data-driven prioritization (Pareto) and root cause analysis (5 Whys, etc.) work best hand-in-hand (ComplianceQuest, n.d.).
We addressed limitations and cautions as well. In practice, IT leaders must ensure that focusing on the frequent issues doesn’t blind them to critical infrequent risks (Purple Griffon, n.d.-b). They should maintain quality data and avoid misusing Pareto by oversimplifying or ignoring the smaller issues entirely (Purple Griffon, n.d.-b). The guidance provided – such as weighting by severity, regularly updating analysis, and using good judgment alongside data – will help avoid those pitfalls. Ultimately, Pareto analysis should serve decision making, not replace it; human insight is needed to interpret and act on the numbers correctly.
Modern ITSM tools and automation have made Pareto analysis easier and more continuous. With dashboard visualizations, real-time analytics, and even AI-driven clustering, IT teams can keep a live pulse on what the top issues are (ServiceNow, n.d.-b; ServiceNow, n.d.-c). This integration into workflows ensures that the moment an issue starts dominating, it’s brought to attention – embodying the proactive stance ITIL advocates (catch it in Problem Management before it causes a major incident). Automation can also take over routine tasks (like generating charts, or even raising problem records), freeing problem managers to focus on analysis and resolution. Embracing these tools is a best practice, but as noted, they work best in an environment of strong process discipline and clear ownership.
We concluded with best practices in governance, data, and KPIs, emphasizing that for Pareto Analysis to drive real improvement, it must be embedded in the ITSM process and culture. Organizations should govern by facts (using data to justify and prioritize problem fixes), maintain reliable data streams (so the facts are trustworthy), and measure outcomes (to ensure the fixes actually deliver expected benefits) (Purple Griffon, n.d.-a). When problem management teams showcase reductions in incidents or faster resolution times as a result of their Pareto-focused initiatives, it reinforces the value of structured problem-solving to the broader business.
In closing, the 80/20 principle in ITIL Problem Management is more than an analysis technique – it’s a mindset of working smarter by addressing what matters most. It directs finite resources to the areas of greatest pain or gain. For ITSM professionals, adopting this mindset means routinely asking: “What should we tackle next that will yield the biggest improvement in stability or efficiency?” Pareto Analysis provides the empirical evidence to answer that question, pointing to, say, a specific service, component, or failure mode. By following that trail and then applying rigorous root cause analysis, teams can eliminate the top problems and measurably elevate the quality of IT services.
The journey of continuous improvement is never truly finished – as improvements are made, new challenges emerge (sometimes the “80/20” rule reappears at a new level of detail). But with a Pareto-guided approach, IT organizations can be confident they are always working on the most impactful opportunities for improvement. This not only maximizes operational performance but also builds credibility with the business, as IT delivers tangible reductions in disruptions and better alignment with business needs.
In essence, Pareto Analysis helps ITIL practitioners cut through noise and focus on value – turning data into insight, and insight into action that drives continuous service excellence. As part of a mature ITSM practice, it ensures that the effort invested in problem management is repaid many times over in the form of fewer incidents, happier users, and more stable, cost-effective IT operations. And that is the ultimate payoff of applying the 80/20 principle in IT: a leaner, smarter, and more proactive IT organization that constantly learns and improves by addressing the causes that truly matter.
BetterExplained. (n.d.). Understanding the Pareto principle (80/20 rule). Retrieved from https://betterexplained.com/articles/understanding-the-pareto-principle-the-8020-rule/
ComplianceQuest. (n.d.). Pros and cons of 5-Why, Pareto & Fishbone diagram. Retrieved from https://www.compliancequest.com/blog/pros-cons-of-5why-pareto-fishbone-diagram/
Float. (n.d.). Pareto’s principle. Retrieved from https://www.float.com/resources/paretos-principle
Giva, Inc. (n.d.). What is the 80/20 rule (Pareto principle) & how does it apply to ITSM? Retrieved from https://www.givainc.com/blog/what-is-80-20-rule-pareto-principle-how-does-it-apply-to-itsm/
Investopedia. (n.d.). 80/20 rule (Pareto principle). Retrieved from https://www.investopedia.com/terms/1/80-20-rule.asp
InvGate. (n.d.). Four problem management root cause analysis techniques explained. Retrieved from https://blog.invgate.com/4-problem-management-root-cause-analysis-techniques-explained
KnowledgeHut. (n.d.). RCA in ITIL service management. Retrieved from https://www.knowledgehut.com/blog/it-service-management/rca-itil
Luxembourg Institute of Science and Technology. (n.d.). TIPA T10 ITIL PAM r2 v4.1 [PDF]. Retrieved from https://www.list.lu/fileadmin/files/projects/TIPA_T10_ITIL_PAM_r2_v4.1.pdf
ManageEngine. (n.d.). Problem management techniques. Retrieved from https://www.manageengine.com/products/service-desk/itsm/problem-management-techniques.html
Medium. (n.d.). Pareto analysis for BI: applying the 80/20 rule visually. Retrieved from https://medium.com/microsoft-power-bi/pareto-analysis-for-bi-applying-the-80-20-rule-visually-0dfe014fced3
Microsoft Power BI Community. (n.d.). Pareto 80/20 calculation for items making up 80 of sales. Retrieved from https://community.fabric.microsoft.com/t5/Desktop/Pareto-80-20-calculation-for-items-making-up-80-of-Sales/m-p/30715
Mobile2B. (n.d.). Root cause analysis tools for effective problem solving. Retrieved from https://www.mobile2b.com/blog/root-cause-analysis-tools-effective-problem-solving
National Institute of Standards and Technology. (2023). Computer security incident handling guide (SP 800-61 Rev. 3). Retrieved from https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-61r3.pdf
National Institute of Standards and Technology. (n.d.). Root cause analysis [Glossary entry]. Retrieved from https://csrc.nist.gov/glossary/term/root_cause_analysis
New Relic. (n.d.). Performing effective root cause analysis. Retrieved from https://newrelic.com/blog/how-to-relic/performing-effective-root-cause-analysis
Purple Griffon. (n.d.-a). ITIL problem management. Retrieved from https://purplegriffon.com/blog/itil-problem-management
Purple Griffon. (n.d.-b). Pareto analysis. Retrieved from https://purplegriffon.com/blog/pareto-analysis
ServiceNow. (n.d.-a). Create DV Pareto visual definition (task). Retrieved from https://www.servicenow.com/docs/bundle/zurich-now-intelligence/page/use/reporting/task/create-dv-pareto-vd.html
ServiceNow. (n.d.-b). Create Pareto charts (Reporting concepts). Retrieved from https://www.servicenow.com/docs/bundle/zurich-now-intelligence/page/use/reporting/concept/c_CreateParetoCharts.html
ServiceNow. (n.d.-c). Pareto report and demand insights (knowledge management). Retrieved from https://www.servicenow.com/docs/bundle/zurich-servicenow-platform/page/product/knowledge-management/concept/pareto-report-demand-insights.html
Tempo. (n.d.). IT problem management. Retrieved from https://www.tempo.io/solutions/itsm/it-problem-management
University of Edinburgh. (2020). Brailsford or Pareto prioritisation and marginal gains. Retrieved from https://blogs.ed.ac.uk/itiltattle/2020/11/06/brailsford-or-pareto-prioritisation-and-marginal-gains/
Zenkins. (n.d.). Root cause analysis in IT support. Retrieved from https://zenkins.com/knowledge-base/root-cause-analysis-in-it-support/
Copyright © 2025 Serhiy Kuzhanov. All rights reserved.
No part of this website may be reproduced, stored in a retrieval system, or transmitted in any form
or by any means without the written permission of the website owner.
