hardmultiple choiceObjective-mapped

Exhibit

Lambda logs:
REPORT RequestId: 9d6b... Duration: 184.27 ms Billed Duration: 185 ms Memory Size: 1024 MB Max Memory Used: 612 MB Init Duration: 812.43 ms

Traffic pattern:
- Low traffic outside weekdays 09:00-09:15 UTC
- Predictable spike every weekday
- Function language: Python 3.12
- No need to keep spare capacity all day

Based on the exhibit, a serverless checkout API is implemented in AWS Lambda and deployed in one Region. The function has a cold-start time of 700-900 ms on the first request after idle periods. Marketing launches a predictable traffic spike every weekday at 09:00 UTC, and the p95 latency target is under 150 ms during the first five minutes of the spike. What should the solutions architect do to meet the latency target while controlling cost?

Question 1hardmultiple choice
Full question →

Based on the exhibit, a serverless checkout API is implemented in AWS Lambda and deployed in one Region. The function has a cold-start time of 700-900 ms on the first request after idle periods. Marketing launches a predictable traffic spike every weekday at 09:00 UTC, and the p95 latency target is under 150 ms during the first five minutes of the spike. What should the solutions architect do to meet the latency target while controlling cost?

Answer choices

Why each option matters

Good practice is not just finding the correct option. The wrong answers often show the exact trap the exam wants you to fall into.

A

Distractor review

Increase the Lambda memory size and leave concurrency at the default value.

More memory can improve execution speed, but it does not eliminate cold starts when a function has been idle.

B

Best answer

Configure provisioned concurrency and scale it up before the predictable spike begins.

Provisioned concurrency keeps pre-initialized environments ready, which removes most cold-start latency. Because the spike is predictable, you can scale concurrency before 09:00 UTC and reduce it afterward to control cost.

C

Distractor review

Put the Lambda function behind an Application Load Balancer so the load balancer absorbs the initialization delay.

An ALB does not remove Lambda initialization time; it only changes how requests are routed to the function.

D

Distractor review

Set reserved concurrency to the expected peak so Lambda will pre-create execution environments.

Reserved concurrency limits concurrent executions, but it does not pre-warm execution environments or eliminate cold starts.

Common exam trap

Common exam trap: answer the scenario, not the keyword

Many certification questions include familiar terms but test a specific constraint. Read the exact wording before choosing an answer that is generally true but wrong for this case.

Technical deep dive

How to think about this question

This question should be treated as a scenario, not a definition check. Identify the problem, the constraint and the best action. Then compare each option against those facts.

KKey Concepts to Remember

  • Read the scenario before looking for a memorised answer.
  • Find the constraint that changes the correct option.
  • Eliminate answers that are true in general but not in this case.
  • Use explanations to understand the rule behind the answer.

TExam Day Tips

  • Underline the problem statement mentally.
  • Watch for words such as best, first, most likely and least administrative effort.
  • Review why wrong options are wrong, not only why the correct option is correct.

Related practice questions

Related SAA-C03 practice-question pages

Use these pages to review the topic behind this question. This is how one missed question becomes focused revision.

More questions from this exam

Keep practising from the same exam bank, or move into a focused topic page if this question exposed a weak area.

FAQ

Questions learners often ask

What does this SAA-C03 question test?

Read the scenario before looking for a memorised answer.

What is the correct answer to this question?

The correct answer is: Configure provisioned concurrency and scale it up before the predictable spike begins. — The exhibit shows that cold-start initialization, not code execution time, is pushing latency above target. Provisioned concurrency is the Lambda feature that keeps environments initialized and ready to serve requests. Because the traffic spike happens on a known schedule, the team can scale provisioned concurrency shortly before the spike and reduce it afterward, which meets the latency target without paying for excess always-on capacity all day. Increasing memory may reduce execution time, but it does not prevent cold starts. An ALB is not a warm-up mechanism for Lambda; it simply forwards requests. Reserved concurrency is a protection and throttling control, not a pre-initialization feature, so it will not solve the first-request delay shown in the logs.

What should I do if I get this SAA-C03 question wrong?

Then try more questions from the same exam bank and focus on understanding why the wrong options are tempting.

Discussion

Loading comments…

Sign in to join the discussion.