70-463 Premium Bundle

70-463 Premium Bundle

Implementing a Data Warehouse with Microsoft SQL Server 2012 Certification Exam

4.5 
(10050 ratings)
0 QuestionsPractice Tests
0 PDFPrint version
November 21, 2024Last update

Microsoft 70-463 Free Practice Questions

Q1. You are developing a SQL Server Integration Services (SSIS) package. 

The package uses a data flow task to source data from a SQL Server database for loading into a dimension table in a data warehouse. 

You need to create a separate data flow path for data that has been modified since it was last processed. 

Which data flow components should you use to identify modified data? (Each correct answer presents a complete solution. Choose all that apply.) 

A. Multicast 

B. Data Conversion 

C. Lookup 

D. Slowly Changing Dimension 

E. Aggregate 

Answer: A,C 

Explanation: A: The transformation that distributes data sets to multiple outputs. 

The transformation that distributes data sets to multiple outputs. 

C: Lookup Transformation 

The transformation that looks up values in a reference table using an exact match. 

Note: 

* SQL Server Integration Services provides three different types of data flow components: 

sources, transformations, and destinations. Sources extract data from data stores such as 

tables and views in relational databases, files, and Analysis Services databases. 

Transformations modify, summarize, and clean data. Destinations load data into data 

stores or create in-memory datasets. 

Incorrect: Not B: Data Conversion Transformation The transformation that converts the data type of a column to a different data type. Not D: Slowly Changing Dimension Transformation The transformation that configures the updating of a slowly changing dimension. Not E: The Aggregate transformation applies aggregate functions, such as Average, to column values and copies the results to the transformation output. Besides aggregate functions, the transformation provides the GROUP BY clause, which you can use to specify groups to aggregate across. 

Q2. You are designing a package control flow. The package moves sales order data from a SQL Azure transactional database to an on-premise reporting database. The package will run several times a day, while new sales orders are being added to the transactional database. The current design of the package control flow is shown in the answer area. (Click the Exhibit button.) 

The Insert New Orders Data Flow task must meet the following requirements: 

Usage of the tempdb database should not be impacted. 

Concurrency should be maximized, while only reading committed transactions. 

If the task fails, only that task needs to be rolled back. 

You need to configure the Insert New Orders Data Flow task to meet the requirements. 

How should you configure the transaction properties? (To answer, select the appropriate setting or settings in the answer area.) 

Select “IsolationLeve as Choas” References: 

http://msdn.microsoft.com/en-us/library/ms137690.aspx http://msdn.microsoft.com/en-us/library/ms137749.aspx http://msdn.microsoft.com/en-us/library/microsoft.sqlserver.dts.runtime.dtscontainer.isolationlevel.aspx http://msdn.microsoft.com/en-us/library/ms173763.aspx 

Answer:  

Q3. You develop a SQL Server Integration Services (SSIS) package that imports SQL Azure data into a data warehouse every night. 

The SQL Azure data contains many misspellings and variations of abbreviations. To import the data, a developer used the Fuzzy Lookup transformation to choose the closest-matching string from a reference table of allowed values. The number of rows in the reference table is very large. 

If no acceptable match is found, the Fuzzy Lookup transformation passes a null value. 

The current setting for the Fuzzy Lookup similarity threshold is 0.50. 

Many values are incorrectly matched. 

You need to ensure that more accurate matches are made by the Fuzzy Lookup transformation without degrading performance. 

What should you do? 

A. Change the Exhaustive property to True. 

B. Change the similarity threshold to 0.55. 

C. Change the similarity threshold to 0.40. 

D. Increase the maximum number of matches per lookup. 

Answer:

Explanation: 

http://msdn.microsoft.com/en-us/library/ms137786.aspx 

Q4. You are adding a new capability to several dozen SQL Server Integration Services (SSIS) packages. 

The new capability is not available as an SSIS task. Each package must be extended with the same new capability. 

You need to add the new capability to all the packages without copying the code between packages. 

What should you do? 

A. Use the Expression task. 

B. Use the Script task. 

C. Develop a custom task. 

D. Use the Script component, 

E. Develop a custom component. 

Answer:

Explanation: References: 

http://msdn.microsoft.com/en-us/library/ms135965.aspx http://msdn.microsoft.com/en-us/library/ms345161.aspx 

Q5. You are designing a data warehouse that contains a customer dimension. 

The customer dimension contains the original customer attributes as well as the most recent set of updated attributes. When a customer is updated, the original customer attributes must be preserved. 

The customer dimension design must: 

Support a Type 3 Slowly Changing Dimension (SCD) process 

Minimize the total storage consumed 

You need to design a solution that meets these requirements. 

What should you do? 

A. Design two tables. One table will hold the original customer surrogate key and attributes. The other table will hold the most recent customer surrogate key and attributes. 

B. Design a table as a Type 2 SCD dimension that holds the surrogate key for each customer and its attributes. Limit the number of records per customer within the table to two. 

C. Design a table to hold the surrogate key for each customer and its attributes. Add a column to store an indicator that identifies the record as original data or recent data. 

D. Design a table to hold the surrogate key for each customer and its attributes as well as additional columns to store the original attributes. 

Answer:

Q6. You are building a SQL Server Integration Services (SSIS) package to load product data sourced from a SQL Azure database to a data warehouse. Before the product data is loaded, you create a batch record by using an Execute SQL task named Create Batch. After successfully loading the product data, you use another Execute SQL task named Set Batch Success to mark the batch as successful. 

You need to create and execute an Execute SQL task to mark the batch as failed if either the Create Batch or Load Products task fails. 

Which three steps should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.) 

Answer:  

Q7. You develop a SQL Server Integration Services (SSIS) project by using the Package Deployment Model. A package in the project extracts data from a Windows Azure SQL Database database. 

The package is deployed to SQL Server. 

The package is not producing the desired results. 

You need to generate the .mdmp and .tmp debug files in order to troubleshoot the issues. 

What should you do? 

A. Execute the catalog.add_data_tap stored procedure with the package execution_id. 

B. Execute the catalog.create_execution_dump stored procedure with the package execution_id. 

C. Run the DTEXEC utility with the /DumpOnError option. 

D. Run the DTEXEC utility with the /Reporting V option. 

Answer:

Q8. You are deploying a new SQL Server Integration Services (SSIS) project to the test environment. 

A package in the project uses a custom task component. 

You need to ensure that the custom object is deployed on the test environment correctly. 

What should you do? 

A. Run the package by using the dtexec /rep /conn command. 

B. Create a reusable custom logging component. 

C. Create an OnError event handler. 

D. Use the gacutil command. 

E. Use the dtutil /copy command. 

F. Deploy the package to the Integration Services catalog by using dtutil and use SQL Server to store the configuration. 

G. Run the package by using the dtexec /dumperror /conn command. 

H. Use the Project Deployment Wizard. 

I. Deploy the package by using an msi file. 

J. Add a data tap on the output of a component in the package data flow. 

K. Run the package by using the dtexecui.exe utility and the SQL Log provider. 

Answer:

Reference: 

http://msdn.microsoft.com/en-us/library/ms403356.aspx 

Q9. You are the administrator of a server that hosts Data Quality Server for a large retail company. 

The server had a hardware failure during business hours. 

You need to restore the server that hosts Data Quality Server to another server. You have a recent backup of all the required databases. 

What should you do? (Each correct answer presents part of the solution. Choose all that apply.) 

A. Restore the DQS_MAIN, DQS_PROJECTS, and DQS_STAGING_DATA databases to another server as soon as possible. 

B. Execute the DQS_MAIN.internal_core.RestoreDQDatabases stored procedure with the appropriate parameter. 

C. Restore only the DQS_MAIN and DQS_STAGING_DATA databases to another server as soon as possible. 

D. Execute the DQS_MAIN.internal_core.InitServer stored procedure with the appropriate parameter. 

Answer: B,D 

Explanation: Steps to restore DQS Databases: 

Restore DQS_MAIN database. 

Restore the DQS_PROJECTS database. 

Restore the DQS_STAGING_DATA database. 

In Object Explorer, right-click the server, and then click.New Query. 

In the Query Editor window, copy the following SQL statements, and 

replace.<PASSWORD>.with the password that you provided during the DQS 

installation for the database master key: 

USE [DQS_MAIN] 

GO 

EXECUTE [internal_core].[RestoreDQDatabases] '<PASSWORD>' 

GO 

.Press F5 to execute the statements. Check the.Results.pane to verify that the statements have executed successfully. 

Note: 

* Backup and restore of SQL Server databases are common operations that database administrators perform for preventing loss of data in a case of disaster by recovering data from the backup databases. Data Quality Server is primarily implemented by two SQL Server databases: DQS_MAIN and DQS_PROJECTS. The backup and restore procedures of the Data Quality Services (DQS) databases are similar to any other SQL Server databases. 

Q10. You are developing a SQL Server Integration Services (SSIS) project that copies a large amount of rows from a SQL Azure database. The project uses the Package Deployment 

Model. This project is deployed to SQL Server on a test server. 

You need to ensure that the project is deployed to the SSIS catalog on the production server. 

What should you do? 

A. Open a command prompt and run the dtexec /dumperror /conn command. 

B. Create a reusable custom logging component and use it in the SSIS project. 

C. Open a command prompt and run the gacutil command. 

D. Add an OnError event handler to the SSIS project. 

E. Open a command prompt and execute the package by using the SQL Log provider and running the dtexecui.exe utility. 

F. Open a command prompt and run the dtexec /rep /conn command. 

G. Open a command prompt and run the dtutil /copy command. 

H. Use an msi file to deploy the package on the server. 

I. Configure the SSIS solution to use the Project Deployment Model. 

J. Configure the output of a component in the package data flow to use a data tap. 

K. Run the dtutil command to deploy the package to the SSIS catalog and store the configuration in SQL Server. 

Answer:

Explanation: References: http://msdn.microsoft.com/en-us/library/hh231102.aspx 

http://msdn.microsoft.com/en-us/library/hh213290.aspx http://msdn.microsoft.com/en-us/library/hh213373.aspx 

Q11. You are developing a SQL Server Integration Services (SSIS) package that imports data into a data warehouse hosted on SQL Azure. 

The package uses a Foreach container to process text files found in a folder. The package must be deployed to a single server by using the Project Deployment model. 

Multiple SQL Server Agent jobs call the package. Each job is executed on a different schedule. Each job passes a different folder path to the package. 

You need to configure the package to accept the folder path from each job. 

Which package configuration should you use? 

A. .dtsConfig file 

B. Registry Entry 

C. Environment Variable 

D. Parent Package Variable 

E. XML Configuration File 

Answer:

Reference: http://msdn.microsoft.com/en-us/library/ms345179.aspx 

http://msdn.microsoft.com/en-us/library/ms141708.aspx http://msdn.microsoft.com/en-us/library/hh213214.aspx http://msdn.microsoft.com/en-us/library/hh213296.aspx http://msdn.microsoft.com/en-us/library/hh213293.aspx 

Q12. You are developing a SQL Server Integration Services (SSIS) project with multiple packages to copy data to a Windows Azure SQL Database database. 

An automated process must validate all related Environment references, parameter data types, package references, and referenced assemblies. The automated process must run on a regular schedule. 

You need to establish the automated validation process by using the least amount of administrative effort. 

What should you do? 

A. Use an event handler for OnError for the package. 

B. Use an event handler for OnError for each data flow task. 

C. Use an event handler for OnTaskFailed for the package. 

D. View the job history for the SQL Server Agent job. 

E. View the All Messages subsection of the All Executions report for the package. 

F. Store the System::SourceID variable in the custom log table. 

G. Store the System::ServerExecutionID variable in the custom log table. 

H. Store the System::ExecutionInstanceGUID variable in the custom log table. 

I. Enable the SSIS log provider for SQL Server for OnError in the package control flow. 

J. Enable the SSIS log provider for SQL Server for OnTaskFailed in the package control flow. 

K. Deploy the project by using dtutil.exe with the /COPY DTS option. 

L. Deploy the project by using dtutil.exe with the /COPY SQL option. 

M. Deploy the .ispac file by using the Integration Services Deployment Wizard. 

N. Create a SQL Server Agent job to execute the SSISDB.catalog.validate_project stored procedure. 

O. Create a SQL Server Agent job to execute the SSISDB.catalog.validate_package stored procedure. 

P. Create a SQL Server Agent job to execute the SSISDB.catalog.create_execution and SSISDB.catalog.start_execution stored procedures. 

Q. Create a table to store error information. Create an error output on each data flow destination that writes OnError event text to the table. 

R. Create a table to store error information. Create an error output on each data flow destination that writes OnTaskFailed event text to the table. 

Answer:

Q13. You are editing a SQL Server Integration Services (SSIS) project named Project1 in SQL Server Data Tools. 

A package Connection Manager has been parameterized with project scope. 

You need to display the parameters that have been generated for the Connection Manager. 

What should you use? To answer, select the appropriate setting or settings in the answer area. 

Answer:  

Q14. You are writing a SQL Server Integration Services (SSIS) package that transfers data from a legacy system. 

Data integrity in the legacy system is very poor. Invalid rows are discarded by the package but must be logged to a CSV file for auditing purposes. 

You need to establish the best technique to log these invalid rows while minimizing the amount of development effort. 

What should you do? 

A. Add a data tap on the output of a component in the package data flow. 

B. Deploy the package by using an msi file. 

C. Run the package by using the dtexecui.exe utility and the SQL Log provider. 

D. uses the dtutil /copy command. 

E. Deploy the package to the Integration Services catalog by using dtutil and use SQL Server to store the configuration. 

F. Create an OnError event handler. 

G. uses the Project Deployment Wizard. 

H. Use the gacutil command. 

I. Create a reusable custom logging component. 

J. Run the package by using the dtexec /rep /conn command. 

K. Run the package by using the dtexec /dumperror /conn command. 

Answer:

Reference: 

http://www.rafael-salas.com/2012/01/ssis-2012-quick-peek-to-data-taps.html 

http://msdn.microsoft.com/en-us/library/hh230989.aspx 

http://msdn.microsoft.com/en-us/library/jj655339.aspx 

Q15. You are designing a SQL Server Integration Services (SSIS) package. The package moves order-related data to a staging table named Order. Every night the staging data is truncated and then all the recent orders from the online store database are inserted into the staging table. 

Your package must meet the following requirements: 

. If the truncate operation fails, the package execution must stop and report an error. . If the Data Flow task that moves the data to the staging table fails, the entire refresh operation must be rolled back. . For auditing purposes, a log entry must be entered in a SQL log table after each execution of the Data Flow task. 

The TransactionOption property for the package is set to Required. 

You need to design the package to meet the requirements. 

How should you design the control flow for the package? (To answer, drag the appropriate setting from the list of settings to the correct location or locations in the answer area.) 

Answer:  

Q16. You are designing an enterprise star schema that will consolidate data from three independent data marts. One of the data marts is hosted on SQL Azure. 

Most of the dimensions have the same structure and content. However, the geography dimension is slightly different in each data mart. 

You need to design a consolidated dimensional structure that will be easy to maintain while ensuring that all dimensional data from the three original solutions is represented. 

What should you do? 

A. Create a junk dimension for the geography dimension. 

B. Implement change data capture. 

C. Create a conformed dimension for the geography dimension. 

D. Create three geography dimensions. 

Answer:

START 70-463 EXAM