Analyze and optimize resources across FinOps Scopes to match actual usage patterns, while ensuring that workloads operate efficiently, sustainably, and generate sufficient business value relative to their cost.
Moving beyond traditional IaaS workloads to include all technology categories, Usage Optimization is a set of practices that ensure resources across all FinOps Scopes are properly selected, correctly sized, only run when needed, appropriately configured, and highly utilized in order to meet all functional and non-functional requirements at the lowest cost and environmental impact. This work is primarily done by Engineering, using guidelines and strategies formed collaboratively with FinOps, Product, and other relevant personas.
Engineers, FinOps Practitioners, and other Personas should seek to ensure there is sufficient business value for the costs associated with each area of usage being consumed. Because systems and services are built and adopted iteratively, it is typical to observe usage patterns and resource utilization over time to ensure performance, availability, and other quality metrics are met, and to adjust or modify resources which are over- or under-sized or underutilized.
There is a strong relationship among all of the Capabilities in the Optimize Usage & Cost Domain. Each Capability in this Domain works in different ways to optimize value — through commitment-based discounts, rearchitecting, managing license and SaaS usage, providing guidance on sustainability improvements, and optimizing the utilization and efficiency of the resources and workloads that make up systems. Among all of these, Usage Optimization will likely be the most widely practiced Capability, with the broadest range of optimization options across all FinOps Scopes.
Early in the FinOps practice, stakeholder Personas will likely play a large role in identifying opportunities to optimize workloads, but over time Engineering will take on the primary responsibility for their usage by seeking out ways to optimize, or better yet by building in optimization as much as possible as systems are being built. No matter how well built and efficient a system is when built, services in the are constantly being added and modernized, and organizations must be prepared to continuously work to keep pace and maintain optimal performance and utilization. Engineering leadership is critical to establishing the cadence and highlighting the need to maintain optimization of workloads at the appropriate level.
A key way the FinOps Practitioners can support optimization activities is by developing a Usage Optimization strategy. This strategy can direct optimization work by highlighting which types of resources and usage areas should be prioritized, setting thresholds for taking action so that time is not wasted on trivial improvements, defining target KPIs the organization wants to achieve, and creating guidelines for making the tradeoffs that come with optimization across all FinOps Scopes.
Other Capabilities in this Domain may have important inputs to this strategy, including highlighting for Engineering where the organization supports (or plans to stop) using licensed software or SaaS, when rearchitecting is preferred over resource optimization, how to prioritize usage optimization against rate optimization, and how to incorporate sustainability and carbon impact considerations into usage optimization decision making. The strategy may also set Leadership’s expectations for how frequently and diligently optimization should be pursued by Engineering relative to new feature development work.
Engineering Personas, in collaboration with FinOps Practitioners, Product, and Leadership Personas, will leverage the Capabilities in the Understand Usage & Cost Domain to review usage and resources across their areas of responsibility. Determining utilization and identifying scaling or usage management opportunities may require access to utilization, performance, or observability data in addition to usage, cost, and carbon impact data. Engineering teams may focus their efforts on finding optimization opportunities in different ways depending on factors such as the criticality of the system, time available to optimize, maturity of the application or service, the nature of the usage type being optimized, or whether the resources are supporting production or non-production environments.
A wide range of options exist to optimize usage across all FinOps Scopes, including:
Examine usage patterns carefully for longer-cycle periods of high utilization (e.g. higher utilization at month-end or during quarterly busy periods) and be cautious of resources that have specific requirements for warranty, contractual, or software performance reasons. Rightsizing typically requires recreating or reconfiguring resources, which can involve service disruptions that should be carefully coordinated with Engineering and other relevant stakeholder personas.
There may be times when utilization needs to remain higher than optimal and the extra expense incurred is justified by the business value the resources create. Equally, the opposite may be true — cost efficiency goals may take precedence and carbon and/or performance expectations can be adjusted accordingly to improve cost outcomes.
For some resource types, such as storage, it may be necessary to estimate latent inefficiency in the stored data, and by extension the potential savings that can be realized by removing or rightsizing that inefficiency. Different data sets require tailored approaches. For example, highly compressible but uncompressed data has relatively high latent inefficiency, whereas encrypted data has relatively low or no latent inefficiency. Data that is infrequently accessed but stored in a high cost, high performance storage tier also has relatively high latent inefficiency. Good data housekeeping practices, such as optimizing data placement, implementing compression techniques, adopting tiered storage solutions, and reducing unnecessary data duplication, can improve both cost efficiency and environmental impact by minimizing the resources required to store and manage data over time.
Technology providers and vendors regularly release modernized offerings, like new generations of compute families, serverless or managed versions of existing services, or new tiers of service with improved price-performance characteristics. Each of these modernization developments should prompt a review of existing usage to identify opportunities to migrate to newer, more cost-efficient and energy-efficient alternatives.
Newer resource types and service tiers typically deliver better cost-performance per unit, and often carry a lower environmental impact as well. Not every modernization opportunity requires immediate action, but Engineering and other FinOps Personas should maintain awareness of new and updated services across all technology categories, and factor modernization into their ongoing optimization planning rather than treating it as a one-time activity.
For any of these decisions to be made, resource utilization, efficiency, sustainability, and cost must be evaluated together and in context. Determining when and where usage optimization can be done effectively involves estimating not only the savings or cost avoidance that can accrue from a change, but also the full cost of making that change, labor hours, service disruptions, and the potential complexity of transforming how a resource or service is consumed in the process.
Create tailored optimization strategies that balance financial, operational, and environmental considerations in a way that generates clear and measurable business value. Include targeted guidance to create context for applying usage optimization activities across key FinOps Scopes and technology categories. For example:
Moving from identifying optimization opportunities that are technically possible to realizing value from those activities requires close alignment between FinOps, Engineering, and other stakeholder Personas. Bridging the gap between opportunity identification and action, agreeing on priorities, timelines, and accountability across all FinOps Scopes is the key focus of this Capability.
Method of objective scoring based on resource cost efficiency.
The Cost Optimization Index score, or COIN, is a quantitative measure designed to assess cloud cost efficiency. COIN applies to any breakdown of infrastructure cost: Team, service, account, etc. COIN is calculated using the savings opportunity and overall total cost for the infrastructure in question to assess efficiency. Think of it as the inverse of waste.
COIN Score = [1 – (Total Savings Opportunity / Total Cost)] * 100
The resulting score from 0-100 serves as an objective benchmark for cost efficiency. Total Cost is based on the aggregate cost of any relevant scope of infrastructure and measured in any desired currency.
Total Savings Opportunity is based on the sum of individual Savings Opportunities. Savings Opportunities are usage patterns within that scope of infrastructure which indicate expected areas of inefficiency and waste. These are calculated as projected savings in the same currency as Total Cost. Each Savings Opportunity will need it’s own cost model to identify potential savings.
This scoring system can help:
Example:
Acme Corp has identified 3 potential areas of savings which they wish to drive through their organization.
Savings Area #1 is defined as the use of older generation storage technologies.
Savings Area #2 is defined as low CPU utilization.
Savings Area #3 is defined as turning on a vendor’s network cost-savings option.
For a given Team Rocket, the FinOps team has calculated the following:
First, calculate the expected savings from each savings opportunity.
Next, calculate the total savings opportunity by summing the individual savings opportunities:
The COIN Score for Team Rocket then is calculated:
Team Rocket’s COIN score reflects that ~12% of their spend is known waste and that overall, based on their aggregate spend and defined patterns of waste, their spend is ~88% efficient.
Data Sources:
Measures the return on investment of optimization initiatives by comparing the combined financial and operational benefits to the cost of implementing the optimization. The formula evaluates whether optimization efforts such as infrastructure right-sizing, efficiency improvements, or performance tuning deliver sufficient value through reduced costs and improved performance relative to the resources required to implement them.
Measures the return on investment of optimization initiatives by comparing the combined financial and operational benefits to the cost of implementing the optimization. The formula evaluates whether optimization efforts such as infrastructure right-sizing, efficiency improvements, or performance tuning deliver sufficient value through reduced costs and improved performance relative to the resources required to implement them. This KPI was developed by the FinOps for Data Center Working Group.
Optimization ROI = (Cost Savings + Performance Gains) / Implementation Cost
Candidate Data Sources:
Measures effectiveness of an auto-scaling system.
Measures effectiveness of an auto-scaling system. The goal of auto-scaling efficiency is to ensure that the right amount of resources are provisioned and de-provisioned in response to changes in demand, in order to achieve a balance between performance, cost, and resource utilization.
Auto-scaling efficiency rate = Maximum capacity cost of running workload to meet workload demand / Cost of running workload with auto-scaling to meet same workload demand. The higher the efficiency rate the more effective the auto-scaling is. Effective Cost can be used in this formula, or the List Cost metric can be used to eliminate the effect of discounts and focus entirely on the scaling effect.
Data Sources: CSP Billing Data
Measures how efficiently a data center uses energy by comparing the total power consumed by the facility to the power consumed by IT equipment. The formula quantifies the overhead required to support IT operations such as cooling, power distribution, and lighting relative to the energy directly used for compute, storage, and networking. A PUE value
Measures how efficiently a data center uses energy by comparing the total power consumed by the facility to the power consumed by IT equipment. The formula quantifies the overhead required to support IT operations such as cooling, power distribution, and lighting relative to the energy directly used for compute, storage, and networking. A PUE value closer to 1.0 indicates higher energy efficiency, meaning a greater proportion of facility power is delivered to IT equipment rather than supporting infrastructure. This KPI was developed by the FinOps for Data Center Working Group.
Power Usage Effectiveness (PUE) = Total Facility Power / IT Equipment Power
Candidate Data Sources:
Optimizing Electricity PUE, WUE, Between Regions, or Cloud Service Providers
Power Usage Effectiveness (PUE) is a metric that evaluates the energy efficiency of a data center by measuring the ratio between the total energy consumed by the data center and the energy used specifically by its IT equipment, such as servers, storage, and network devices. Similarly, Water Usage Effectiveness (WUE) quantifies water usage efficiency in a data center by calculating the ratio of total water usage, measured in liters, to the total IT energy consumption, measured in kilowatt-hours (kWh).
PUE = Total Facility Energy / IT Equiment Energy
WUE = Total Water Used by the Data Center / Total IT Equipment Energy Consumption (in kWh)
Data Sources:
Additional Guidance: