Spring Sale 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: save70

MuleSoft-Platform-Architect-I Exam Dumps : Salesforce Certified MuleSoft Platform Architect (Mule-Arch-201)

PDF
MuleSoft-Platform-Architect-I pdf
 Real Exam Questions and Answer
 Last Update: May 12, 2026
 Question and Answers: 152 With Explanation
 Compatible with all Devices
 Printable Format
 100% Pass Guaranteed
$27  $90
MuleSoft-Platform-Architect-I exam
PDF + Testing Engine
MuleSoft-Platform-Architect-I PDF + engine
 Both PDF & Practice Software
 Last Update: May 12, 2026
 Question and Answers: 152
 Discount Offer
 Download Free Demo
 24/7 Customer Support
$42  $140
Testing Engine
MuleSoft-Platform-Architect-I Engine
 Desktop Based Application
 Last Update: May 12, 2026
 Question and Answers: 152
 Create Multiple Test Sets
 Questions Regularly Updated
  90 Days Free Updates
  Windows and Mac Compatible
$31.5  $105

Verified By IT Certified Experts

CertsTopics.com Certified Safe Files

Up-To-Date Exam Study Material

99.5% High Success Pass Rate

100% Accurate Answers

Instant Downloads

Exam Questions And Answers PDF

Try Demo Before You Buy

Certification Exams with Helpful Questions And Answers

Salesforce Certified MuleSoft Platform Architect (Mule-Arch-201) Questions and Answers

Question 1

A new upstream API Is being designed to offer an SLA of 500 ms median and 800 ms maximum (99th percentile) response time. The corresponding API implementation needs to sequentially invoke 3 downstream APIs of very similar complexity.

The first of these downstream APIs offers the following SLA for its response time: median: 100 ms, 80th percentile: 500 ms, 95th percentile: 1000 ms.

If possible, how can a timeout be set in the upstream API for the invocation of the first downstream API to meet the new upstream API's desired SLA?

Options:

A.

Set a timeout of 50 ms; this times out more invocations of that API but gives additional room for retries

B.

Set a timeout of 100 ms; that leaves 400 ms for the other two downstream APIs to complete

C.

No timeout is possible to meet the upstream API's desired SLA; a different SLA must be negotiated with the first downstream API or invoke an alternative API

D.

Do not set a timeout; the Invocation of this API Is mandatory and so we must wait until it responds

Buy Now
Question 2

An organization has created an API-led architecture that uses various API layers to integrate mobile clients with a backend system. The backend system consists of a number of specialized components and can be accessed via a REST API. The process and experience APIs share the same bounded-context model that is different from the backend data model. What additional canonical models, bounded-context models, or anti-corruption layers are best added to this architecture to help process data consumed from the backend system?

Options:

A.

Create a bounded-context model for every layer and overlap them when the boundary contexts overlap, letting API developers know about the differences between upstream and downstream data models

B.

Create a canonical model that combines the backend and API-led models to simplify and unify data models, and minimize data transformations.

C.

Create a bounded-context model for the system layer to closely match the backend data model, and add an anti-corruption layer to let the different bounded contexts cooperate across the system and process layers

D.

Create an anti-corruption layer for every API to perform transformation for every data model to match each other, and let data simply travel between APIs to avoid the complexity and overhead of building canonical models

Question 3

A Rate Limiting policy is applied to an API implementation to protect the back-end system. Recently, there have been surges in demand that cause some API client

POST requests to the API implementation to be rejected with policy-related errors, causing delays and complications to the API clients.

How should the API policies that are applied to the API implementation be changed to reduce the frequency of errors returned to API clients, while still protecting the back-end

system?

Options:

A.

Keep the Rate Limiting policy and add 9 Client ID Enforcement policy

B.

Remove the Rate Limiting policy and add an HTTP Caching policy

C.

Remove the Rate Limiting policy and add a Spike Control policy

D.

Keep the Rate Limiting policy and add an SLA-based Spike Control policy