Pre-Summer Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code = getmirror

Pass the Salesforce Application Architect Data-Architect Questions and answers with ExamsMirror

Practice at least 50% of the questions to maximize your chances of passing.
Exam Data-Architect Premium Access

View all detail and faqs for the Data-Architect exam


793 Students Passed

95% Average Score

96% Same Questions
Viewing page 2 out of 8 pages
Viewing questions 11-20 out of questions
Questions # 11:

Northern Trail Outfitter has implemented Salesforce for its associates nationwide, Senior management is concerned that the executive dashboard is not reliable for their real-time decision-making. On analysis, the team the following issues with data entered in Salesforce.

Information in certain records is incomplete.

Incorrect entry in certain fields causes records to be excluded in report fitters.

Duplicate entries cause incorrect counts.

Which three steps should a data architect recommend to address the issues?

Options:

A.

Periodically export data to cleanse data and import them back into Salesforce for executive reports.

B.

Build a sales data warehouse with purpose-build data marts for dashboards and senior management reporting.

C.

Explore third-party data providers to enrich and augment information entered in salesforce.

D.

Leverage Salesforce features, such as validate rules, to avoid incomplete and incorrect records.

E.

design and implement data-quality dashboard to monitor and act on records that are incomplete or incorrect

Questions # 12:

Universal Containers has successfully migrated 50 million records into five different objects multiple times in a full copy sandbox. The Integration Engineer wants to re-run the test again a month before it goes live into Production. What is the recommended approach to re-run the test?

Options:

A.

Truncate all 5 objects quickly and re-run the data migration test.

B.

Refresh the full copy sandbox and re-run the data migration test.

C.

Hard delete all 5 objects’ data and re-run the data migration test.

D.

Truncate all 5 objects and hard delete before running the migration test.

Questions # 13:

UC has large amount of orders coming in from its online portal. Historically all order are assigned to a generic user.

Which 2 measures should data architect recommend to avoid any performance issues while working with large number of order records? Choose 2 answers:

Options:

A.

Clear the role field in the generic user record.

B.

Salesforce handles the assignment of orders automatically and there is no performance impact.

C.

Create a role at top of role hierarchy and assign the role to the generic user.

D.

Create a pool of generic users and distribute the assignment of memory to the pool of users.

Questions # 14:

A Customer is migrating 10 million order and 30 million order lines into Salesforce using Bulk API. The Engineer is experiencing time-out errors or long delays querying parents order IDs in Salesforce before importing related order line items. What is the recommended solution?

Options:

A.

Query only indexed ID field values on the imported order to import related order lines.

B.

Leverage an External ID from source system orders to import related order lines.

C.

Leverage Batch Apex to update order ID on related order lines after import.

D.

Leverage a sequence of numbers on the imported orders to import related order lines.

Questions # 15:

Universal Containers has a legacy system that captures Conferences and Venues. These Conferences can occur at any Venue. They create hundreds of thousands of Conferences per year. Historically, they have only used 20 Venues. Which two things should the data architect consider when denormalizing this data model into a single Conference object with a Venue picklist? Choose 2 answers

Options:

A.

Limitations on master -detail relationships.

B.

Org data storage limitations.

C.

Bulk API limitations on picklist fields.

D.

Standard list view in -line editing.

Questions # 16:

Universal Containers is exporting 40 million Account records from Salesforce using Informatica Cloud. The ETL tool fails and the query log indicates a full table scan time-out failure. What is the recommended solution?

Options:

A.

Modify the export job header to specify Export-in-Parallel.

B.

Modify the export job header to specify Sforce-Enable-PKChunking.

C.

Modify the export query that includes standard index fields(s).

D.

Modify the export query with LIMIT clause with Batch size 10,000.

Questions # 17:

UC has a classic encryption for Custom fields and is leveraging weekly data reports for data backups. During the data validation of exported data UC discovered that encrypted field values are still being exported as part of data exported. What should a data architect recommend to make sure decrypted values are exported during data export?

Options:

A.

Set a standard profile for Data Migration user, and assign view encrypted data

B.

Create another field to copy data from encrypted field and use this field in export

C.

Leverage Apex class to decrypt data before exporting it.

D.

Set up a custom profile for data migration user and assign view encrypted data.

Questions # 18:

What should a data architect do to provide additional guidance for users when they enter information in a standard field?

Options:

A.

Provide custom help text under field properties.

B.

Create a custom page with help text for user guidance.

C.

Add custom help text in default value for the field.

D.

Add a label field with help text adjacent to the custom field.

Questions # 19:

Universal Containers (UC) has a multi-level master-detail relationship for opportunities, a custom opportunity line item object, and a custom discount request. UC has opportunity as master and custom line item object as detail in master-detail relationship. UC also has a custom line item object as master and a custom discount request object as detail in another master-detail relationship. UC has a requirement to show all sums of discounts across line items at an opportunity level. What is the recommended solution to address these requirements?

Options:

A.

Use roll-up for the line-item-level summary and a trigger for the opportunity amount summary, as only one level roll-up is allowed.

B.

Update the master-detail relationships to lookup relationships in order to allow the discount amount to roll up.

C.

Remove the master-detail relationships and rely completely on workflow/triggers to summarize the discount amount.

D.

Roll-up discount request amount at the line-item-level and line-item-level summary discount at the opportunity level.

Questions # 20:

US has released a new disaster recovery (DR)policy that states that cloud solutions need a business continuity plan in place separate from the cloud providers built in data recovery solution.

Which solution should a data architect use to comply with the DR policy?

Options:

A.

Leverage a 3rd party tool that extract salesforce data/metadata and stores the information in an external protected system.

B.

Leverage salesforce weekly exports, and store data in Flat files on a protected system.

C.

Utilize an ETL tool to migrate data to an on-premise archive solution.

D.

Write a custom batch job to extract data changes nightly, and store in an external protected system.

Viewing page 2 out of 8 pages
Viewing questions 11-20 out of questions
TOP CODES

TOP CODES

Top selling exam codes in the certification world, popular, in demand and updated to help you pass on the first try.