70-768 Exam - Developing SQL Data Models (beta)

certleader.com

Q1. HOTSPOT - (Topic 3)

A database named DB2 uses the InMemory query mode. Users frequently run the following query:

You need to configure SQL Server Profiler to determine why the query is performing poorly. Which three event should you monitor on the SQL Server Profiler trace events

configuration page? To answer, select the appropriate options in the answer area.

Answer:

Explanation:

By using SQL Profiler, you can intercept two classes of trace events from Analysis Services, DAX Query Plan and DirectQuery events, both generated by the DirectQuery engine. Here, in this scenario we have a DAX Query.

DAX Query Plan events are generated by the DAX formula.

By using the In-Memory mode, you store a copy of data in the xVelocity (VertiPaq) storage engine.

Figure: This is how a query is executed by using In-Memory mode.

References: Microsoft SQL Server 2012 Analysis Services, The BISM Tabular Model,

Microsoft Press (July 2012), page 331

From Scenario: Users report that the query takes a long time to complete.

Topic 4, Mix Questions

10.  - (Topic 4)

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution. Determine whether the solution meets the stated goals.

You have an existing multidimensional cube that provides sales analysis. The users can slice by date, product, location, customer, and employee.

The management team plans to evaluate sales employee performance relative to sales targets. You identify the following metrics for employees:

You need to implement the KPI based on the Status expression. Solution: You design the following solution:

Does the solution meet the goal?

A. Yes

B. No

Q2. - (Topic 1)

You need to configure the server to optimize the afternoon report generation based on the OrderAnalysis cube.

Which property should you configure?

A. LowMemoryLimit

B. VertiPaqPagingPolicy

C. TotalMemoryLimit

D. VirtualMemoryLimit

Answer: A

Explanation:

LowMemoryLimit: For multidimensional instances, a lower threshold at which the server first begins releasing memory allocated to infrequently used objects.

From scenario: Reports that are generated based on data from the OrderAnalysis cube take more time to complete when they are generated in the afternoon each day. You examine the server and observe that it is under significant memory pressure.

Q3. - (Topic 1)

You need to configure the server to optimize the afternoon report generation based on the OrderAnalysis cube.

Which property should you configure?

A. LowMemoryLimit

B. VertiPaqPagingPolicy

C. TotalMemoryLimit

D. VirtualMemoryLimit

Answer: A

Explanation:

LowMemoryLimit: For multidimensional instances, a lower threshold at which the server first begins releasing memory allocated to infrequently used objects.

From scenario: Reports that are generated based on data from the OrderAnalysis cube take more time to complete when they are generated in the afternoon each day. You examine the server and observe that it is under significant memory pressure.

Q4.  - (Topic 4)

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution. Determine whether the solution meets the stated goals.

You deploy a tabular data model to an instance of Microsoft SQL Server Analysis Services (SSAS). The model uses an in-memory cache to store and query data. The data set is already the same size as the available RAM on the server. Data volumes are likely to continue to increase rapidly.

Your data model contains multiple calculated tables.

The data model must begin processing each day at 2:00 and processing should be complete by 4:00 the same day. You observe that the data processing operation often does not complete before 7:00. This is adversely affecting team members.

You need to improve the performance.

Solution: Change the storage mode for the data model to DirectQuery. Does the solution meet the goal?

A. Yes

B. No

Answer: A

Explanation:

By default, tabular models use an in-memory cache to store and query data. When tabular models query data residing in-memory, even complex queries can be incredibly fast. However, there are some limitations to using cached data. Namely, large data sets can exceed available memory, and data freshness requirements can be difficult if not impossible to achieve on a regular processing schedule.

DirectQuery overcomes these limitations while also leveraging RDBMS features making query execution more efficient.

With DirectQuery: +

Data is up-to-date, and there is no extra management overhead of having to maintain a separate copy of the data (in the in-memory cache). Changes to the underlying source data can be immediately reflected in queries against the data model.

Datasets can be larger than the memory capacity of an Analysis Services server. Etc.

References:https://docs.microsoft.com/en-us/sql/analysis-services/tabular-models/directquery-mode-ssas-tabular

Q5.  - (Topic 4)

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution. Determine whether the solution meets the stated goals.

You deploy a tabular data model to an instance of Microsoft SQL Server Analysis Services (SSAS). The model uses an in-memory cache to store and query data. The data set is already the same size as the available RAM on the server. Data volumes are likely to continue to increase rapidly.

Your data model contains multiple calculated tables.

The data model must begin processing each day at 2:00 and processing should be complete by 4:00 the same day. You observe that the data processing operation often does not complete before 7:00. This is adversely affecting team members.

You need to improve the performance. Solution: Enable Buffer Cache Extensions. Does the solution meet the goal?

A. Yes

B. No

Answer: B

Explanation:

In this scenario we would need both Buffer Cache Extensions and SSD.

The buffer pool extension provides the seamless integration of a nonvolatile random access memory (that is, solid-state drive) extension to the Database Engine buffer pool to

significantly improve I/O throughput.

References:https://docs.microsoft.com/en-us/sql/database-engine/configure-windows/buffer-pool-extension

Q6. DRAG DROP - (Topic 2)

You need to configure the SalesAnalysis cube to correct the sales analysis by customer calculation.

Which four actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

Answer:

Explanation:

Step 1: Open the cube editor, and open the Dimension Usage tab.

Step 2: Configure a relationship between the Customer dimension and the Sales measure group. Use Day as the granularity.

From scenario: The SalesAnalysis cube contains a fact table named CoffeeSale loaded from a table named FactSale in the data warehouse. The time granularity within the cube is 15 minutes. The cube is processed every night at 23:00. You determine that the fact table cannot be fully processed in the expected time. Users have reported slow query response times.

Step 3: Reprocess the cube.

Step 4: Deploy the project changes.

Q7. HOTSPOT - (Topic 4)

You are a database administrator in a company that uses Microsoft SharePoint Server for all intranet sites. You are responsible for the installation of new database server instances.

You must install Microsoft SQL Server Analysis Server (SSAS) to support deployment of the following projects. You develop both projects by using SQL Server Data Tools.

You need to install the appropriate services to support both projects.

What should you do? In the table below, select the appropriate services for each project. NOTE: Make only one selection in each column. Each correct selection is worth one point.

Answer:

Explanation:

Project1:

Project2: Multidimensional

Note: Analysis Services can be installed in one of three server modes: Multidimensional and Data Mining (default), Power Pivot for SharePoint, and Tabular.

Q8. HOTSPOT - (Topic 2)

You need to configure the project option settings to minimize deployment time for the CustomerAnalysis data model.

What should you do? To answer, select the appropriate setting from each list in the answer area.

Answer:

Explanation:

Scenario:

Box 1, Processing option:Default

Process Default detects the process state of database objects, and performs processing necessary to deliver unprocessed or partially processed objects to a fully processed state. If you change a data binding, Process Default will do a Process Full on the affected object.

Note: Processing Method This setting controls whether the deployed objects are processed after deployment and the type of processing that will be performed. There are three processing options:

Default processing (default) Full processing

None

Box 2, Transactional deployment: False

If this option is False, Analysis Services deploys the metadata changes in a single transaction, and deploys each processing command in its own transaction.

Scenario: The CustomerAnalysis data model will contain a large amount of data and needs to be shared with other developers even if a deployment fails. Each time you deploy a change during development, processing takes a long time.

References:https://docs.microsoft.com/en-us/sql/analysis-services/multidimensional- models/deployment-script-files-specifying-processing-options

Topic 3, Case Study #3

Background

You are a developer for a Seattle-based company. The company is expanding globally. Many company employees speak fluent Mandarin and read Simplified Chinese.

You have six tabular data models that are deployed to two instances of Microsoft SQL Server Analysis Services (SSAS).

Users report that the query takes a long time to complete.

You are planning the disk space allocations for a new Microsoft SQL Server Analysis Services deployment. You plan to move several relational data file databases to the new SSAS instance. The databases require a total of 10 GB of disk space.

You also plan to deploy Cubes and Aggregations and use Object Processing. Cubes will have small fact tables and few dimension members. No unnecessary aggregations will be created. You plan to process an entire cube in a single transaction.

Data Models

One of the data models is named CustomerSales. This data model contains eight tables. The model includes a table named Sales that defines several measures, including a measure named PriorYearSales. The PriorYearSales measure is referenced by other measures, and is not intended to be analyzed directly by users. You must translate the metadata for all fata the CustomerSales data model to Simplifies Chinese. Team members from the Shanghai office assist with identifying appropriate translations.

A data model named OrderAnalysis is deployed to one of the SSAS instances. Order data is loaded into the OrderAnalysis data as part of an overnight process. You observe that the model is not up-to-date.

The business analysis team uses a variety of client applications to issue MDX queries against OrderAnalysis. Order data must be completely up-to-date.

The OrderAnalysis model has two user-defined hierarchies that are defined in a table named Order. New customers are only added once per day. The overnight process is sufficiently up-to-date for the Customer data to provide optimal performance while achieving the data currency goals whenever possible.

Databases

You deploy a database named DB1 to an SSAS instance as a project by using SQL Server Data Tools. Data analysts report that they cannot access near real time data from the SSAS SalesAnalysis model from DB1. You discover that the project has been deployed with the Direct Query Mode option set to OFF.

Most queries that use the SalesAnalysis data model use data from a table named FactInternetSales that is 20 gigabyte (GB) in size. Cached data must be available for the FactInternetSales table. All queries accessing the SalesAnalysis model must be executed in near real time.

Q9.  - (Topic 4)

Note: This question is part of a series of questions that use the same or similar answer choices. An answer choice may be correct for more than one question in the series. Each question is independent of the other questions in this series. Information and details provided in a question apply only to that question.

You have a Microsoft SQL Server Analysis Services (SSAS) instance that is configured to use multidimensional mode. You create the following cube:

Users need to be able to analyze sales by color.

You need to create a dimension that contains all of the colors for products sold by the company.

Which relationship type should you use between the InternetSales table and the new dimension?

A. no relationship

B. regular

C. fact

D. referenced

E. many-to-many

F. data mining

Answer: B

Explanation:

A regular dimension relationship between a cube dimension and a measure group exists when the key column for the dimension is joined directly to the fact table.

References: https://docs.microsoft.com/en-us/sql/analysis-services/multidimensional-models-olap-logical-cube-objects/dimension-relationships

Q10.  - (Topic 4)

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution. Determine whether the solution meets the stated goals.

You have an existing multidimensional cube that provides sales analysis. The users can slice by date, product, location, customer, and employee.

The management team plans to evaluate sales employee performance relative to sales targets. You identify the following metrics for employees:

You need to implement the KPI based on the Status expression. Solution: You design the following solution:

Does the solution meet the goal?

A. Yes

B. No

Answer: B