Elastic Compute in F&O: From Fixed Tiers to Shared Capacity

For a more simple overview see at the bottom of this blog, that also includes a calculator.

Microsoft is changing how infrastructure works in F&O apps. In unified environments, elastic compute replaces fixed environment tiers with a shared-capacity model that automatically scales based on actual usage. Instead of choosing a Tier 2, Tier 3, or similar size up front, sandbox and production environments draw from a common Power Platform capacity model.

In practical terms, this means customers no longer plan around fixed infrastructure sizes in the same way. The AOS tier remains the core compute layer, but the number of AOS instances can increase as demand grows. Microsoft describes this as a model where compute scales with real workload rather than with a preselected environment tier.

Microsoft’s published reference points are important. Each AOS instance aligns with 650,000 Power Platform Requests, or PPRs. Sandbox and production environments have a minimum of two AOS instances, and an environment can scale up to 80 AOS in total, split across up to 40 interactive and 40 batch instances. Unified Developer Environments are the exception and stay fixed at one AOS.

Why this is simpler

The big shift is that capacity is no longer something you buy as a named infrastructure tier. Instead, capacity comes from a tenant-wide PPR entitlement. Microsoft defines that entitlement as 500,000 base PPRs, plus 5,000 PPRs for each assigned user license, with optional 50,000-PPR add-on packs when more headroom is needed.

That makes the model easier to explain to customers. The old question was, “Which environment tier do we need?” The new question is closer to, “How much workload do we expect, and how much shared capacity do we want available?” That is a simpler model operationally, but it also means customers need to think more about usage patterns across the tenant.

Cost follows workload

PPRs are not presented by Microsoft as a direct per-request price list, but Microsoft does sell 50,000-PPR add-on packs. That gives customers a practical reference point for understanding what extra headroom may cost, even though the exact commercial price can vary by region and agreement.

The important change is conceptual. Under the old model, customers often paid for fixed capacity whether they used it fully or not. Under elastic compute, idle baseline capacity becomes less of a planning problem, while heavy activity becomes much more visible. High concurrency, API bursts, Dataverse-driven traffic, and large batch workloads are now the things that matter most.

There is also no rollover concept here. Power Platform request limits are measured over a rolling 24-hour period, so unused request headroom is not something customers can bank and carry forward into a later spike. That reinforces the idea that this is a live consumption model rather than a stored monthly allowance.

The most important thing to understand: 5,000 and 40,000 are not the same number

This is where many people get confused. Microsoft uses one model for elastic-compute capacity and another for licensed-user request limits. For elastic compute, the number is 5,000 PPR per assigned user license, added to the tenant-wide base of 500,000. That number helps determine how far environments can scale.

Separately, Microsoft documents licensed-user Power Platform request limits. For paid Dynamics 365 Enterprise users, the documented request limit is 40,000 requests per user per 24 hours in the request-limits model. That number is relevant for request-limit behavior and throttling, but it should not be added into the elastic-compute AOS calculation. These are two different mechanisms answering two different questions.

A useful way to explain it is this. The 5,000 figure helps answer, “How much elastic-compute capacity does this license contribute to the tenant?” The 40,000 figure helps answer, “How many Power Platform requests can this licensed user make in a 24-hour period before request limits become relevant?” They are related, but they are not interchangeable.

Example: 100 Supply Chain Management users

Take a customer with 100 assigned Supply Chain Management users. For elastic compute, Microsoft’s formula gives 500,000 base PPR plus 100 × 5,000 PPR, which equals 1,000,000 PPR. Microsoft’s own example table maps that to 2 AOS of capacity.

So the way customers should think about this is not, “100 users means 4,000,000 PPR for scaling.” That would be mixing the user request-limit model with the environment capacity model. For elastic compute, the relevant number is the 1,000,000-PPR tenant entitlement, not the 40,000-per-user request-limit figure.

It is also important to remember that attach licenses do not create extra platform entitlements. Microsoft’s licensing guide states that attach licenses do not include additional platform entitlements and instead use the entitlements of the assigned base license. So customers should not assume that a base license plus an attach license doubles their elastic-compute accrual.

Multiple environments still matter

Even though sandbox and production now use the same elastic-compute model, customers still need to think at tenant level. Test, UAT, Golden, and other environments are no longer planned as isolated fixed tiers. They sit within the same overall capacity model, and operational demand across them can still matter when testing, refreshes, migrations, integrations, and batch processing happen in parallel.

This becomes especially relevant during implementation. At that stage, customers often have several active environments before the final production licensing position is fully established. If those environments need to scale during migration, test cycles, or parallel workload periods, extra 50,000-PPR add-on packs may need to be purchased temporarily until the final licensing model is in place.

What gets cheaper, and what gets harder

The part that likely becomes cheaper is unused infrastructure. Customers no longer have to overbuy fixed tiers just to be safe. Baseline environments and idle capacity become less wasteful, and sandbox-versus-production planning becomes simpler because both follow the same capacity model.

The part that can become more expensive is sustained activity. Heavy integrations, API traffic, Dataverse calls, and large batch workloads now translate more directly into visible capacity demand. The model is more efficient, but it also makes cost more sensitive to actual usage.

That also makes budgeting a bit harder. Under fixed tiers, infrastructure cost was easier to predict even if it was inefficient. Under elastic compute, customers may need to adjust PPR capacity over time, especially in projects with fluctuating demand or several active environments. The model is cleaner technically, but less static financially.

So if you don’t have more PPR capacity available and want to add a second sandbox environment, expect to buy 40 PPR add-on packs at approx 50 USD each = approx 2.000 USD/month.

And if you need additional DB capacity remember to pick up some Dataverse Database Capacity add-on at 40 USD/GB/Month.

A few important limits to keep in mind

Not all compute is fully covered by this model today. Components such as Commerce Scale Unit and eCommerce-related workloads are not clearly described as part of the same elastic-compute model in the public documentation, so customers should be careful not to assume full alignment there yet. This is still a moving area.

It is also worth noting that unified environments are managed through the Power Platform Admin Center, not through Lifecycle Services. Microsoft’s current documentation explicitly positions finance and operations unified environments as PPAC-managed with zero LCS footprint.

Final thoughts

Elastic compute is a meaningful shift for F&O. It replaces fixed tiers with a shared, PPR-driven capacity model, gives sandbox and production the same compute framework, and allows customers to add capacity without going through the old infrastructure process.

The simplest customer takeaway is this: stop thinking in tiers, start thinking in workload. If the workload is light, the model is simpler and often more efficient. If the workload is heavy or unpredictable, the model is still flexible, but it requires closer attention to capacity, integrations, and budgeting.

PS: AI helped with parts of this post, but I’m the author.

References :
Licencing information : https://www.microsoft.com/en-us/licensing/product-licensing/dynamics365
Microsoft documentation on elastic compute with F&O: https://learn.microsoft.com/en-us/power-platform/admin/unified-experience/elastic-compute




And I got this from a friend (if he want’s to share his name, he can comment 🙂 ):


Elastic Compute for Finance & Operations Apps
Microsoft Power Platform · Unified Environments

Elastic Compute

Finance & Operations apps automatically scale compute power based on actual usage — no fixed tiers, no support tickets, no guesswork.

Why Elastic Compute?

Three fundamental advantages over the legacy Lifecycle Services model.

Same Performance Model

Sandbox and production environments share the same elastic compute model. Performance testing in sandbox reflects what you’ll see in production.

No Tier Selection

Compute capacity is determined by your Power Platform Request (PPR) entitlement and scales automatically — no tier lock-in.

Simple Capacity Adds

Purchase more PPR add-on packs from the M365 admin center to increase your compute ceiling instantly — no contracts or escalations.

At a Glance

80
Max AOS Instances per Environment
40 + 40
Interactive + Batch AOS Split
650K
PPRs per AOS Instance
500K
Tenant-wide Base PPRs

How Elastic Compute Works

The AOS tier is the primary compute component. Instances are stateful, hold session state in memory, and are fronted by Azure load balancers with session affinity.

Architecture Overview

Users 👤👤👤 Interactive Sessions OData / API Calls Power Platform Dataverse Virtual Tables Flows & Plug-ins Batch Jobs Background Processing Azure Load Balancer ⚖️ Session Affinity Request Distribution Routes to same AOS per user session AOS Interactive #1 X++ Business Logic Session State in Memory AOS Interactive #2 X++ Business Logic Session State in Memory … up to 40 interactive AOS Batch #1–40 Background Job Processing Independent Scaling Persistent Storage Layer 🗄️ Azure SQL Metadata & Transactional State Caching Layer Performance Acceleration 📦 Blob Storage Documents & Binary Data Externalized — survives scale events
Key Insight: AOS instances are stateful — they hold user session state in memory. This is why scale-up (adding instances) is seamless, but scale-down (removing instances) is reserved for maintenance windows to avoid session loss.

Deploy to the Middle

Environments start at a mid-sized baseline — not minimum, not maximum — so they’re ready for typical workloads immediately.

Baseline & Elastic Scaling Model

Maximum (80 AOS) Elastic Scale-Up Zone Automatic · No downtime · Transparent Baseline Deployment (“the Middle”) Below Baseline Environment never scales below baseline Minimum (2 AOS) Peak demand Normal load Scale-down only during maintenance windows Compute Capacity Time →

Scale-Up Triggers

👥 High Concurrent Sessions

Many interactive users accessing the environment simultaneously drives scale-up of interactive AOS instances.

🔗 Burst API Traffic

Spikes in OData or custom service calls trigger additional AOS instances to absorb the load.

Heavy Batch Processing

Large batch job queues trigger independent scale-up of batch AOS instances.

🔄 Power Platform Calls

Increased Dataverse virtual-table traffic from flows, plug-ins, or apps triggers elastic scale-up.

Scale-Up & Scale-Down Flow

1

Demand Detected

Platform monitors sessions, API traffic & batch queue depth

2

Scale Up

New AOS instances added horizontally — zero downtime

3

Load Redistributed

Load balancer distributes traffic across all instances

4

Maintenance Window

Sessions drained, extra AOS removed, returns to baseline

Power Platform Requests & Compute Capacity

PPRs are the unit of measure governing how much compute is allocated. Each AOS instance requires 650,000 PPRs.

How You Accrue PPRs

Tenant-wide Base

500,000 PPRs

Automatically included with any Dynamics 365 purchase.

Per-User License

5,000 PPRs / license

Each assigned user license adds to the tenant’s total PPR pool.

Add-on Packs

50,000 PPRs / pack

Purchased from the Microsoft 365 admin center.

PPR Calculation Formula

500,000 Tenant Base PPRs + Licenses × 5,000 Per-User Accrual + Packs × 50,000 Add-on Packs (optional) = Total PPRs ÷ 650,000 = AOS Floor: 2 AOS  •  Cap: 80 AOS (40 interactive + 40 batch)

AOS Capacity by Number of User Licenses

2
20 users
2
100
2
250
4
500
8
1,000
20
2,500
39
5,000
77
10,000+

Each AOS requires 650,000 PPRs  •  Minimum 2 AOS  •  Maximum 80 AOS

User Licenses PPR Calculation Total PPRs AOS Capacity
20 (min)500,000 + (20 × 5,000)600,0002 AOS (floor)
100500,000 + (100 × 5,000)1,000,0002 AOS
250500,000 + (250 × 5,000)1,750,0002 AOS
500500,000 + (500 × 5,000)3,000,0004 AOS
1,000500,000 + (1,000 × 5,000)5,500,0008 AOS
2,500500,000 + (2,500 × 5,000)13,000,00020 AOS
5,000500,000 + (5,000 × 5,000)25,500,00039 AOS
10,000+500,000 + (10,000 × 5,000)50,500,00077 AOS

🧮 PPR & AOS Calculator

500,000
Base PPRs
2,500,000
License PPRs
0
Add-on PPRs
3,000,000
Total PPRs
4
AOS Instances

PPRs Serve Two Purposes

1. Capacity Entitlement

PPRs determine how many AOS instances your environments can scale to. More PPRs = higher elastic compute ceiling.

2. Request Throttling Limit

PPRs also govern throttling limits for Power Platform operations — Dataverse calls, plug-ins, and flows. These limits are independent of AOS capacity.

Unified Developer Environments

Developer environments are an exception to elastic scaling.

Environment Type Comparison

Sandbox & Production Min 2 AOS — Max 80 AOS ⬆️ Scale Up ⚖️ Baseline ⬇️ Scale Down 🔄 Elastic PPR ✓ Full elastic compute Developer (UDE) Fixed 1 AOS — No scaling Single AOS Instance VS Debugger attached ✗ No elastic compute
Why? The Visual Studio debugger must attach to a specific AOS process. A single AOS instance ensures stable debugging. For multi-AOS workloads, use a sandbox environment instead.

Lifecycle Services vs. Unified Environments

Elastic compute eliminates the friction of fixed tiers, manual scaling, and sandbox/production mismatches.

❌ Lifecycle Services (Legacy)

  • Fixed tiers (Tier 2–5) selected at purchase time
  • Sandbox ≠ Production performance
  • Need contract amendments (EA/CSP) to change capacity
  • Support tickets required to scale
  • Separate LCS project for each production environment
  • Max 3–12 AOS per environment (by tier)
  • Cloud-hosted VMs for developers

✅ Unified Environments (Elastic)

  • Elastic compute scales automatically based on PPRs
  • Sandbox = Production — same capacity model
  • Purchase PPR add-ons from M365 admin center
  • Automatic scale-up with zero downtime
  • Provision directly from tenant capacity pool
  • Up to 80 AOS per environment (40 + 40)
  • Unified Developer Environments with single AOS
Aspect Lifecycle Services Unified Environments
Compute ModelFixed tiers at purchaseElastic, auto-scales on PPR
Sandbox = Production?No — different tiers/SQL/AOSYes — same model & max scale
Changing CapacityEA/CSP amendments or ticketsPPR add-ons via M365 admin
Scaling BehaviorManual tier change + downtimeAuto scale-up (no downtime)
Creating EnvironmentsPurchase add-on slots per tierAvailable storage (1 GB needed)
Multiple ProductionsNew LCS project requiredProvision from capacity pool
Perf Testing ReliabilityTier mismatch = unreliableSame compute ceiling
Max AOS / Environment3–12 AOS (varies by tier)Up to 80 AOS (40+40)
Developer EnvironmentsCloud-hosted Azure VMsUDE with single AOS

Increasing Capacity: Add-on Pack Example

A customer with 20 licenses (600K PPRs) purchases 40 add-on packs to reach 2.6M PPRs and go from 2 to 4 AOS.

Add-on Pack Boost

Before 20 licenses → 600K PPRs Below 650K threshold AOS AOS 2 AOS (floor) +40 packs (+2M PPRs) After 600K + 2M = 2.6M PPRs → 4 AOS AOS AOS AOS AOS 4 AOS (2 interactive + 2 batch)

Key Takeaways

Elastic compute replaces fixed tiers with a flexible, PPR-driven model that scales automatically.

Zero
Downtime on Scale-Up
Same
Sandbox & Prod Model
No
Tickets or Amendments
80
Max AOS per Env

Leave a Reply