AWS Lambda cost breakdown for 2026

Équipe d'experts Wiz
Main takeaways from this article:
  • Lambda’s core costs stem from requests and compute, but hidden drivers, such as logs, storage, and networking, can quietly push bills higher.

  • The “always free” tier (1M requests + 400K GB-seconds/month) is often enough for small or event-driven workloads.

  • You can lower Lambda spend by right-sizing, choosing Graviton instead of x86, upping your performance efficiency, and using cost monitoring and management tools.

  • Cost spikes often trace back to misconfigurations—over‑permissive logs, noisy retries, or NAT egress where private endpoints would do. A code‑to‑cloud view that correlates IAM, network paths, log volumes, and function configs helps teams spot the ‘why’ behind spend and fix it at the source.

With adoption on the rise, more companies are looking to control Lamba spend. After all, even small inefficiencies become costly: A single poorly tuned workload can add hundreds to your monthly bill. The upside? Well-structured workloads can unlock the same scale in savings, which is why it’s critical to understand Lambda’s pricing levers.

In this article, we’ll break down the ins and outs of cost management in AWS Lambda. We’ll explore the complexities hidden behind the simple pay-as-you-go model and highlight the overlooked elements that can quietly inflate your bill so that you can optimize with confidence.

2025 Gartner® Market Guide for CNAPP

Security teams are consolidating tools, aligning workflows, and prioritizing platforms that offer end-to-end context. The 2025 Gartner® Market Guide for Cloud-Native Application Protection Platforms (CNAPP) explores this shift and outlines what security leaders should consider as the market matures.

AWS Lambda pricing tiers

As we’ve seen, AWS Lambda’s pricing is designed to be simple and pay-as-you-go. You’re charged based on three main factors:

  1. The number of requests your functions handle

  2. The computing time they consume

  3. The processor architecture 

But it’s worth highlighting the AWS Lambda “always free” tier before we dive into the paid model—in some cases, it can provide everything your project requires.

The AWS Lambda “always free” tier

AWS’s “always free” offerings apply to both new and existing accounts, indefinitely. For Lambda, here’s what you get free each month:

These limits reset every month for as long as you use Lambda, and if your application runs only a modest number of functions, you might never need a paid plan. 

For example, a serverless API endpoint with 256 MB of memory, 1,000,000 requests per month, and an average billed time of 200 ms per invocation (cold start + code invocation) would use less than 400,000 GB-seconds of compute time per month and can remain within the free tier:

Compute time = 0.256 (GB) x 0.2 (s) x 1,000,000 = 51,200 (GB-s)

That said, additional services may add usage-based costs. Some of them also have free-tier offerings, which can help keep costs low as the project scales.

Request-based pricing

Once you pass the free-tier limit, AWS Lambda request pricing typically ranges from $0.20 to $0.28 per additional million requests (as of October 6, 2025), depending on the region. If the app processes 3 million requests in a month in a North American region:

  • The first million is free (free tier)

  • The remaining 2 million are billed at $0.20 per million, which equals $0.40 in request charges

For many small-scale or event-driven applications, request charges are minimal, and it’s compute time that might drive costs.

On-demand compute-based pricing

AWS uses a GB-seconds model to calculate compute time:

Compute Time (GB-seconds) = Allocated Memory (GB) × Total Execution Time (Seconds)

Allocated memory is a fixed setting you choose (128 MB–10,240 MB) that directly impacts both speed and cost. 

The total execution time includes both the INIT phase (cold start) and the active run of the code, and this full duration is what gets billed. Since that duration is multiplied by the number of requests, even small inefficiencies can scale into significant costs.

Figure 1: Lambda execution time

Compute-based pricing varies by region, total GB-seconds used, and system architecture. Per-GB-second rates are fixed per region and architecture. There is no tiered discounting for on-demand compute duration; pricing for Provisioned Concurrency uses separate (often lower) per-GB-second rates.

To see how compute-based pricing works in practice, let’s use the Lambda Pricing Calculator to look at a sample workload running in North America on x86 architecture:

Figure 2: Sample pricing calculation in US East (N. Virginia) (Adapted from AWS)

Architecture-based pricing

AWS Lambda supports two processor architectures: x86_64 (Intel/AMD) and Arm64 powered by AWS Graviton2. While both deliver the same functionality, their pricing and performance can differ:

  • x86_64: Standard baseline GB-second pricing rate; performance varies by workload and runtime

  • Arm64 (Graviton2): Priced lower per GB-second than x86_64 in many regions; AWS reports up to 34% better price/performance and up to 19% better performance vs. x86 for some workloads

AWS Lambda core pricing drivers

Beyond the three primary cost drivers (request charges, compute duration, and architecture), several other factors can affect your Lambda bill.

1. Provisioned Concurrency

Provisioned Concurrency keeps Lambda functions pre-initialized so they can respond in milliseconds, reducing cold start latency for performance-critical applications. You choose how much concurrency to reserve, and AWS keeps that number of environments ready at all times.

In Provisioned Concurrency, charges involve:

  • Configured concurrency × allocated memory × time enabled (rounded up to the nearest 5 minutes)

  • Provisioned Concurrency execution duration uses separate per‑GB‑second rates (often lower than on‑demand); confirm current regional rates on AWS

If traffic exceeds the configured concurrency, excess executions are billed at normal on-demand rates. The free tier does not apply to Provisioned Concurrency.

Figure 3: Provisioned Concurrency price in US East (N. Virginia) (Source: AWS)

2. CloudWatch Logs

When a Lambda function generates output, it can be sent to Amazon CloudWatch Logs if logging permissions are enabled. Until recently, log ingestion was billed at a flat rate, but AWS now uses tiered pricing; the higher the log volume sent, the lower the per-GB rate becomes.

In CloudWatch Logs, charges apply for:

  • Ingestion: Starts at $0.50/GB, decreasing to $0.05/GB at high volumes

  • Storage: Per GB-month for retained logs

  • Querying: CloudWatch Logs Insights is billed per GB scanned

Figure 4: CloudWatch Logs price in US East (N. Virginia) (Source: AWS)

The Lambda free tier doesn’t cover logs. CloudWatch Logs includes a free tier (e.g., 5 GB/month of ingestion and 5 GB/month of archived storage). These logs can also be routed to services like Amazon S3 or Data Firehose for flexibility or lower storage costs.

3. Data transfer

Ingress is free. Egress to the internet or another region follows EC2 data-transfer rates (includes a free tier; regional rates vary).

Figure 5: Standard data transfer price in US East (N. Virginia) (Source: AWS)

4. Mandatory VPC placement overhead

When a Lambda function runs inside an Amazon VPC, AWS automatically creates elastic network interfaces (ENIs) in the subnets to connect to other resources. The ENIs themselves are free, but outbound internet access from a VPC requires a NAT gateway, which adds costs. Additional charges may also apply for VPC peering or VPC endpoints if used:

  • NAT gateway: Charged per hour (e.g., $0.045/hour in US East - N. Virginia), plus per GB of data processed

  • VPC peering: No hourly fee. Within the same region, data processed by the peering connection is billed per GB in each direction (rate varies; see https://aws.amazon.com/vpc/pricing/). Cross‑region transfers follow EC2 data‑transfer rates.

  • VPC endpoints: These offer private access to AWS services. Interface endpoints are billed hourly + per GB, while gateway endpoints (S3/DynamoDB) have no extra charge.

Figure 6: VPC NAT gateway pricing in US East (N. Virginia) (Source: AWS)

5. Configuration errors and recursive invocations

Excessive memory allocation or overly aggressive retry policies can drive up costs without improving performance.

Recursive invocation is another pitfall, where a function triggers itself (e.g., writing to an Amazon SQS queue that calls the same function). This can quickly inflate invocations. AWS now detects most loops after about 16 iterations, but even short ones can still generate costs before being stopped.

AWS Lambda cost-optimization strategies

Managing Lambda costs isn’t just about keeping usage low. It’s about making smart decisions in how functions are configured, monitored, and scaled.

1. Right-sizing

Memory allocation in Lambda directly controls both the RAM and the number of vCPUs assigned to a function. More memory means more compute power, which can shorten execution time for memory or CPU-bound workloads.

In some cases, increasing memory actually lowers total cost, since the function completes faster and uses fewer billed milliseconds. The goal is to balance the two, which is where the AWS Lambda Power Tuning tool helps. By running test payloads across different memory configurations, it maps out the trade-offs between duration, cost, and performance. 

For example, a function running at 512 MB might finish in half the time compared to 256 MB, making it cheaper overall despite the higher memory allocation:

Figure 7: AWS Lambda Power Tuning tool result (Source: AWS)

AWS Compute Optimizer can analyze historical usage data and recommend optimal memory configurations for Lambda. This can be useful when manual tuning experiments aren’t practical for every function.

2. Graviton adoption

AWS Graviton2 processors (Arm64) offer up to 19% better performance at 20% lower cost compared to x86, making them one of the simplest ways to optimize Lambda costs.

In practice, switching to Graviton2 can shorten function duration as well as lower per-GB-second pricing, so you save twice: faster runs and lower unit costs. Many workloads require only a configuration change to run on Graviton2, though some functions may need repackaging if they use x86-specific binaries or dependencies.

3. Performance efficiency

The longer a function runs, the higher the cost and the greater the latency for end users. To increase performance efficiency, you can:

  • Minimize package size so functions download and initialize faster.

  • Reduce dependency complexity by using lighter frameworks.

  • Reuse resources by setting up SDK clients, DB connections, and local /tmp caches outside the handler.

  • Follow language-specific best practices for performance.

4. Architectural patterns for cost efficiency

Selecting the right architectural pattern ensures that costs apply only to real work, not wasted cycles, while also improving scalability and reliability.

  • Event-driven workflows: Use triggers from S3, DynamoDB Streams, or EventBridge so functions execute only in response to real activity. This eliminates the need for constant polling, which can waste invocations.

  • Step Functions orchestration: Break complex tasks into smaller steps. Step Functions manages waiting, retries, and sequencing, so Lambdas aren’t billed while idle.

  • Fan-out/fan-in: Distribute work in parallel using SNS + SQS or Kinesis, then aggregate results efficiently by batching multiple records per invocation, using queue-driven backpressure to auto-scale only when needed, and reducing idle time in the reducer. This improves throughput while preventing bottlenecks and excess invocations.

  • Prevent before deploy: Add guardrails in CI/CD to block functions with excessive logging, missing retries/DLQs, or public egress paths. You’ll avoid both risk and surprise bills.

5. Cost monitoring and management tools

Manually tracking Lambda costs isn’t realistic as workloads grow. Instead, you’ll want to rely on AWS, in-house, or third-party tools for cost visibility, automation, and governance.

AWS offers services including Cost Explorer, Budgets, and Compute Optimizer, but they lack security context. A better bet? Wiz can correlate spend with configuration and security signals to spot anomalies, surface risky misconfigurations, and pinpoint fixes that cut both risk and cost.

6. Compute Savings Plans apply to Lambda

Compute Savings Plans can cover Lambda request and duration charges. If you have steady Lambda usage, committing to a $/hour spend can materially lower effective rates. Model scenarios in the AWS Pricing Calculator and track utilization in Cost Explorer.

7. Ephemeral storage beyond the default

Lambda supports up to 10 GB of /tmp ephemeral storage. Allocations above the default have separate pricing based on GB‑seconds used. Confirm current rates at AWS and include this in cost estimates for storage‑heavy functions.

How Wiz helps optimize Lambda security and cost visibility

Even though Lambda pricing looks straightforward at first, it often relies on multiple AWS and third-party services, which can quickly add hidden costs. Wiz helps organizations get clarity by combining deep visibility into Lambda environments with cost visibility and insights, empowering teams to optimize with confidence.

Wiz gives you unified cost visibility across major cloud providers to help you manage and optimize costs across categories and resource types. By unifying cloud infrastructure context, cost context, utilization context, private pricing context, Wiz surfaces accurate cost optimization opportunities that translate into real savings. Request a demo today to see how Wiz can help you identify the highest priority savings opportunities in your cloud.

See for yourself...

Learn what makes Wiz the platform to enable your cloud security operation

Pour plus d’informations sur la façon dont Wiz traite vos données personnelles, veuillez consulter notre Politique de confidentialité.