Cyber Monday Sale Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 65pass65

MuleSoft-Integration-Architect-I Salesforce Certified MuleSoft Platform Integration Architect (Mule-Arch-202) is now Stable and With Pass Result | Test Your Knowledge for Free

MuleSoft-Integration-Architect-I Practice Questions

Salesforce Certified MuleSoft Platform Integration Architect (Mule-Arch-202)

Last Update 9 hours ago
Total Questions : 273

Dive into our fully updated and stable MuleSoft-Integration-Architect-I practice test platform, featuring all the latest Salesforce MuleSoft exam questions added this week. Our preparation tool is more than just a Salesforce study aid; it's a strategic advantage.

Our Salesforce MuleSoft practice questions crafted to reflect the domains and difficulty of the actual exam. The detailed rationales explain the 'why' behind each answer, reinforcing key concepts about MuleSoft-Integration-Architect-I. Use this test to pinpoint which areas you need to focus your study on.

MuleSoft-Integration-Architect-I PDF

MuleSoft-Integration-Architect-I PDF (Printable)
$43.75
$124.99

MuleSoft-Integration-Architect-I Testing Engine

MuleSoft-Integration-Architect-I PDF (Printable)
$50.75
$144.99

MuleSoft-Integration-Architect-I PDF + Testing Engine

MuleSoft-Integration-Architect-I PDF (Printable)
$63.7
$181.99
Question # 1

The ABC company has an Anypoint Runtime Fabric on VMs/Bare Metal (RTF-VM) appliance installed on its own customer-hosted AWS infrastructure.

Mule applications are deployed to this RTF-VM appliance. As part of the company standards, the Mule application logs must be forwarded to an external log management tool (LMT).

Given the company's current setup and requirements, what is the most idiomatic (used for its intended purpose) way to send Mule application logs to the external LMT?

Options:

A.  

In RTF-VM, install and configure the external LTM's log-forwarding agent

B.  

In RTF-VM, edit the pod configuration to automatically install and configure an Anypoint Monitoring agent

C.  

In each Mule application, configure custom Log4j settings

D.  

In RTF-VM. configure the out-of-the-box external log forwarder

Discussion 0
Question # 2

An organization uses one specific CloudHub (AWS) region for all CloudHub deployments. How are CloudHub workers assigned to availability zones (AZs) when the organization's Mule applications are deployed to CloudHub in that region?

Options:

A.  

Workers belonging to a given environment are assigned to the same AZ within that region.

B.  

AZs are selected as part of the Mule application's deployment configuration.

C.  

Workers are randomly distributed across available AZs within that region.

D.  

An AZ is randomly selected for a Mule application, and all the Mule application's CloudHub workers are assigned to that one AZ

Discussion 0
Question # 3

A large life sciences customer plans to use the Mule Tracing module with the Mapped Diagnostic Context (MDC) logging operations to enrich logging in its Mule application and to improve tracking by providing more context in the Mule application logs. The customer also wants to improve throughput and lower the message processing latency in its Mule application flows.

After installing the Mule Tracing module in the Mule application, how should logging be performed in flows in Mule applications, and what should be changed In the log4j2.xml files?

Options:

A.  

In the flows, add Mule Tracing module Set logging variable operations before any Core Logger components.In log4j2.xml files, change the appender's pattern layout to use %MDC and then assign the appender to a Logger or Root element.

B.  

In the flows, add Mule Tracing module Set logging variable operations before any CoreLogger components.In log4j2.xmI files, change the appender’s pattern layout to use the %MDC placeholder and then assign the appender to an AsyncLogger element.

C.  

In the flows, add Mule Tracing module Set logging variable operations before any CoreLogger components.In log4j2.xmI files, change the appender’'s pattern layout to use %asyncLogger placeholder and then assign the appender to an AsyncLogger element.

D.  

In the flows, wrap Logger components in Async scopes. In log4j2.xmI files, change the appender's pattern layout to use the %asyncLoggerplaceholder and then assign the appender to a Logger or Root element.

Discussion 0
Question # 4

Refer to the exhibit.

A Mule application is deployed to a multi-node Mule runtime cluster. The Mule application uses the competing consumer pattern among its cluster replicas to receive JMS messages from a JMS queue. To process each received JMS message, the following steps are performed in a flow:

Step l: The JMS Correlation ID header is read from the received JMS message.

Step 2: The Mule application invokes an idempotent SOAP webservice over HTTPS, passing the JMS Correlation ID as one parameter in the SOAP request.

Step 3: The response from the SOAP webservice also returns the same JMS Correlation I

D.  

Step 4: The JMS Correlation ID received from the SOAP webservice is validated to be identical to the JMS Correlation ID received in Step 1.

Step 5: The Mule application creates a response JMS message, setting the JMS Correlation ID message header to the validated JMS Correlation ID and publishes that message to a response JMS queue.

Where should the Mule application store the JMS Correlation ID values received in Step 1 and Step 3 so that the validation in Step 4 can be performed, while also making the overall Mule application highly available, fault-tolerant, performant, and maintainable?

Options:

A.  

Both Correlation ID values should be stored in a persistent object store

B.  

Both Correlation ID values should be stored In a non-persistent object store

C.  

The Correlation ID value in Step 1 should be stored in a persistent object storeThe Correlation ID value in step 3 should be stored as a Mule event variable/attribute

D.  

Both Correlation ID values should be stored as Mule event variable/attribute

Discussion 0
Question # 5

An integration Mule application is deployed to a customer-hosted multi-node Mule 4 runtime duster. The Mule application uses a Listener operation of a JMS connector to receive incoming messages from a JMS queue.

How are the messages consumed by the Mule application?

Options:

A.  

Depending on the JMS provider's configuration, either all messages are consumed by ONLY the primary cluster node or else ALL messages are consumed by ALL cluster nodes

B.  

Regardless of the Listener operation configuration, all messages are consumed by ALL cluster nodes

C.  

Depending on the Listener operation configuration, either all messages are consumed by ONLY the primary cluster node or else EACH message is consumed by ANY ONE cluster node

D.  

Regardless of the Listener operation configuration, all messages are consumed by ONLY the primary cluster node

Discussion 0
Question # 6

An organization currently uses a multi-node Mule runtime deployment model within their datacenter, so each Mule runtime hosts several Mule applications. The organization is planning to transition to a deployment model based on Docker containers in a Kubernetes cluster. The organization has already created a standard Docker image containing a Mule runtime and all required dependencies (including a JVM), but excluding the Mule application itself.

What is an expected outcome of this transition to container-based Mule application deployments?

Options:

A.  

Required redesign of Mule applications to follow microservice architecture principles

B.  

Required migration to the Docker and Kubernetes-based Anypoint Platform - Private Cloud Edition

C.  

Required change to the URL endpoints used by clients to send requests to the Mule applications

D.  

Guaranteed consistency of execution environments across all deployments of a Mule application

Discussion 0
Question # 7

An integration Mute application is being designed to process orders by submitting them to a backend system for offline processing. Each order will be received by the Mute application through an HTTPS POST and must be acknowledged immediately. Once acknowledged, the order will be submitted to a backend system. Orders that cannot be successfully submitted due to rejections from the backend system will need to be processed manually (outside the backend system).

The Mule application will be deployed to a customer-hosted runtime and is able to use an existing ActiveMQ broker if needed.

The backend system has a track record of unreliability both due to minor network connectivity issues and longer outages.

What idiomatic (used for their intended purposes) combination of Mule application components and ActiveMQ queues are required to ensure automatic submission of orders to the backend system, while minimizing manual order processing?

Options:

A.  

An On Error scope Non-persistent VM ActiveMQ Dead Letter Queue for manual processing

B.  

An On Error scope MuleSoft Object Store ActiveMQ Dead Letter Queue for manual processing

C.  

Until Successful component MuleSoft Object Store ActiveMQ is NOT needed or used

D.  

Until Successful component ActiveMQ long retry Queue ActiveMQ Dead Letter Queue for manual processing

Discussion 0
Question # 8

An insurance provider is implementing Anypoint platform to manage its application infrastructure and is using the customer hosted runtime for its business due to certain financial requirements it must meet. It has built a number of synchronous API's and is currently hosting these on a mule runtime on one server

These applications make use of a number of components including heavy use of object stores and VM queues.

Business has grown rapidly in the last year and the insurance provider is starting to receive reports of reliability issues from its applications.

The DevOps team indicates that the API's are currently handling too many requests and this is over loading the server. The team has also mentioned that there is a significant downtime when the server is down for maintenance.

As an integration architect, which option would you suggest to mitigate these issues?

Options:

A.  

Add a load balancer and add additional servers in a server group configuration

B.  

Add a load balancer and add additional servers in a cluster configuration

C.  

Increase physical specifications of server CPU memory and network

D.  

Change applications by use an event-driven model

Discussion 0
Question # 9

A leading bank implementing new mule API.

The purpose of API to fetch the customer account balances from the backend application and display them on the online platform the online banking platform. The online banking platform will send an array of accounts to Mule API get the account balances.

As a part of the processing the Mule API needs to insert the data into the database for auditing purposes and this process should not have any performance related implications on the account balance retrieval flow

How should this requirement be implemented to achieve better throughput?

Options:

A.  

Implement the Async scope fetch the data from the backend application and to insert records in the Audit database

B.  

Implement a for each scope to fetch the data from the back-end application and to insert records into the Audit database

C.  

Implement a try-catch scope to fetch the data from the back-end application and use the Async scope to insert records into the Audit database

D.  

Implement parallel for each scope to fetch the data from the backend application and use Async scope to insert the records into the Audit database

Discussion 0
Question # 10

A system API EmployeeSAPI is used to fetch employee's data from an underlying SQL database.

The architect must design a caching strategy to query the database only when there is an update to the employees stable or else return a cached response in order to minimize the number of redundant transactions being handled by the database.

What must the architect do to achieve the caching objective?

Options:

A.  

Use an On Table Row on employees table and call invalidate cacheUse an object store caching strategy and expiration interval to empty

B.  

Use a Scheduler with a fixed frequency every hour triggering an invalidate cache flowUse an object store caching strategy and expiration interval to empty

C.  

Use a Scheduler with a fixed frequency every hour triggering an invalidate cache flowUse an object store caching strategy and set expiration interval to 1-hour

D.  

Use an on table rule on employees table call invalidate cache and said new employees data to cacheUse an object store caching strategy and set expiration interval to 1-hour

Discussion 0
Get MuleSoft-Integration-Architect-I dumps and pass your exam in 24 hours!

Free Exams Sample Questions