Cyber Monday Sale Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmas50

Salesforce MuleSoft-Integration-Architect-I - Salesforce Certified MuleSoft Platform Integration Architect (Mule-Arch-202)

An organization needs to enable access to their customer data from both a mobile app and a web application, which each need access to common fields as well as certain unique fields. The data is available partially in a database and partially in a 3rd-party CRM system. What APIs should be created to best fit these design requirements?

A.

A Process API that contains the data required by both the web and mobile apps, allowing these applications to invoke it directly and access the data they need thereby providing the flexibility to add more fields in the future without needing API changes.

B.

One set of APIs (Experience API, Process API, and System API) for the web app, and another set for the mobile app.

C.

Separate Experience APIs for the mobile and web app, but a common Process API that invokes separate System APIs created for the database and CRM system

D.

A common Experience API used by both the web and mobile apps, but separate Process APIs for the web and mobile apps that interact with the database and the CRM System.

What limits if a particular Anypoint Platform user can discover an asset in Anypoint Exchange?

A.

Design Center and RAML were both used to create the asset

B.

The existence of a public Anypoint Exchange portal to which the asset has been published

C.

The type of the asset in Anypoint Exchange

D.

The business groups to which the user belongs

A customer wants to use the mapped diagnostic context (MDC) and logging variables to enrich its logging and improve tracking by providing more context in the logs.

The customer also wants to improve the throughput and lower the latency of message processing.

As an Mulesoft integration architect can you advise, what should the customer implement to meet these requirements?

A.

Use synchronous logging and use pattern layout with [%MDC] in the log4j2.xml configuration file and then configure the logging variables

B.

Useasync logger at the level greater than INFO and use pattern layout with [%MDC] in the log4j2,xml configuration file and then configure the logging variables

C.

Useasync logger at the level equal to DEBUG orTRACEand use patternlayout with [%MDC] in the log4j2.xml configuration file and then configure the logging variables

D.

Use synchronous logging at the INFO DEBUG or Trace level and use pattern layout with [%MDC] in the log4j2.xml configuration file and then configure the logging variables

What comparison is true about a CloudHub Dedicated Load Balancer (DLB) vs. the CloudHub Shared Load Balancer (SLB)?

A.

Only a DLB allows the configuration of a custom TLS server certificate

B.

Only the SLB can forward HTTP traffic to the VPC-internal ports of the CloudHub workers

C.

Both a DLB and the SLB allow the configuration of access control via IP whitelists

D.

Both a DLB and the SLB implement load balancing by sending HTTP requests to workers with the lowest workloads

One of the backend systems involved by the API implementation enforces rate limits on the number of request a particle client can make.

Both the back-end system and API implementation are deployed to several non-production environments including the staging environment and to a particular production environment. Rate limiting of the back-end system applies to all non-production environments.

The production environment however does not have any rate limiting.

What is the cost-effective approach to conduct performance test of the API implementation in the non-production staging environment?

A.

Including logic within the API implementation that bypasses in locations of the back-end system in the staging environment and invoke a Mocking service that replicates typical back-end system responsesThen conduct performance test using this API implementation

B.

Use MUnit to simulate standard responses from the back-end system.Then conduct performance test to identify other bottlenecks in the system

C.

Create a Mocking service that replicates the back-end system'sproduction performance characteristicsThen configure the API implementation to use the mockingservice and conduct the performance test

D.

Conduct scaled-down performance tests in the staging environment against rate-limiting back-end system. Then upscale performance results to full production scale

A company is modernizing its legal systems lo accelerate access lo applications and data while supporting the adoption of new technologies. The key to achieving this business goal is unlocking the companies' key systems and dala including microservices miming under Docker and kubernetes containers using apis.

Considering the current aggressive backlog and project delivery requirements the company wants to take a strategic approach in the first phase of its transformation projects by quickly deploying API's in mule runtime that are able lo scale, connect to on premises systems and migrate as needed.

Which runtime deployment option supports company's goals?

A.

Customer hosted self provisioned runtimes

B.

Cloudhub runtimes

C.

Runtime fabric on self managed Kubernetes

D.

Runtime fabric on Vmware metal

A leading bank implementing new mule API.

The purpose of API to fetch the customer account balances from the backend application and display them on the online platform the online banking platform. The online banking platform will send an array of accounts to Mule API get the account balances.

As a part of the processing the Mule API needs to insert the data into the database for auditing purposes and this process should not have any performance related implications on the account balance retrieval flow

How should this requirement be implemented to achieve better throughput?

A.

Implement the Async scope fetch the data from the backend application and to insert records in the Audit database

B.

Implement a for each scope to fetch the data from the back-end application and to insert records into the Audit database

C.

Implement a try-catch scope to fetch the data from the back-end application and use the Async scope to insert records into the Audit database

D.

Implement parallel for each scope to fetch the data from the backend application and use Async scope to insert the records into the Audit database

A company is designing a mule application to consume batch data from a partner's ftps server The data files have been compressed and then digitally signed using PGP.

What inputs are required for the application to securely consumed these files?

A.

ATLS context Key Store requiring the private key and certificate for the companyPGP public key of partnerPGP private key for the company

B.

ATLS context first store containing a public certificate for partner ftps server and the PGP public key of the partnerTLS contact Key Store containing the FTP credentials

C.

TLS context trust or containing a public certificate for the ftps server The FTP username and password The PGP public key of the partner

D.

The PGP public key of the partnerThe PGP private key for the companyThe FTP username and password

The AnyAirline organization's passenger reservations center is designing an integration solution that combines invocations of three different System APIs (bookFlight, bookHotel, and bookCar) in a business transaction. Each System API makes calls to a single database.

The entire business transaction must be rolled back when at least one of the APIs fails.

What is the most idiomatic (used for its intended purpose) way to integrate these APIs in near real-time that provides the best balance of consistency, performance, and reliability?

A.

Implement eXtended Architecture (XA) transactions between the API implementationsCoordinate between the API implementations using a Saga patternImplement caching in each API implementation to improve performance

B.

Implement local transactions within each API implementationConfigure each API implementation to also participate in the same eXtended Architecture (XA) transactionImplement caching in each API implementation to improve performance

C.

Implement local transactions in each API implementationCoordinate between the API implementations using a Saga patternApply various compensating actions depending on where a failure occurs

D.

Implement an eXtended Architecture (XA) transaction manager in a Mule application using a Saga patternConnect each API implementation with the Mule application using XA transactionsApply various compensating actions depending on where a failure occurs

A mule application must periodically process a large dataset which varies from 6 GB lo 8 GB from a back-end database and write transform data lo an FTPS server using a properly configured bad job scope.

The performance requirements of an application are approved to run in the cloud hub 0.2 vCore with 8 GB storage capacity and currency requirements are met.

How can the high rate of records be effectively managed in this application?

A.

Use streaming with a file storage repeatable strategy for reading records from the database and batch aggregator with streaming to write to FTPS

B.

Use streaming with an in-memory reputable store strategy for reading records from the database and batch aggregator with streaming to write to FTPS

C.

Use streaming with a file store repeatable strategy for reading records from the database and batch aggregator with an optimal size

D.

Use streaming with a file store repeatable strategy reading records from the database and batch aggregator without any required configuration