Q1. DRAG DROP
You are loading a dataset into SQL Server. The dataset contains numerous duplicates for the Artist and Song columns.
The values in the Artist column in the dataset must exactly match the values in the Artist domain in the knowledge base. The values in the Song column in the dataset can be a close match with the values in the Song domain.
You need to use SQL Server Data Quality Services (DQS) to define a matching policy rule to identify duplicates.
How should you configure the Rule Editor? (To answer, drag the appropriate answers to the answer area.)
Answer:
Q2. A SQL Server Integration Services (SSIS) package was deployed two weeks ago with the Project Deployment Model.
Sometimes the package is started as part of a multistep SQL job. At other times, the package is started manually by a database administrator by using the Object Explorer in SQL Server Management Studio.
You need to identify the authenticated user responsible for starting the package each time it executes.
How can you find this information?
A. In the SSISDB.[catalog], query the .[executions] view.
B. In the SSISDB.[catalog] , query the [event_messages] view.
C. In SQL Server Management Studio, view the SQL Agent Job History.
D. In SQL Server Management Studio, view the SQL Agent Error Log.
E. In SQL Server Management Studio, view the SQL Server Log.
Answer: A
Q3. You are deploying a new SQL Server Integration Services (SSIS) package to several servers.
The package must meet the following requirements:
•.NET Common Language Runtime (CLR) integration in SQL Server must not be enabled.
•The Connection Managers used in the package must be configurable without editing the package.
•The deployment procedure must be automated as much as possible. You need to set up a deployment strategy that meets the requirements. What should you do?
A. Use the gacutil command.
B. Use the dtutil /copy command.
C. Use the Project Deployment Wizard.
D. Create an OnError event handler.
E. Create a reusable custom logging component.
F. Run the package by using the dtexec /rep /conn command.
G. Run the package by using the dtexec /dumperror /conn command.
H. Run the package by using the dtexecui.exe utility and the SQL Log provider.
I. Add a data tap on the output of a component in the package data flow.
J. Deploy the package by using an msi file.
K. Deploy the package to the Integration Services catalog by using dtutil and use SQL Server to store the configuration.
Answer: B
Q4. DRAG DROP
You are validating whether a SQL Server Integration Services (SSIS) package named Master.dtsx in the SSIS catalog is executing correctly.
You need to display the number of rows in each buffer passed between each data flow component of the package.
Which three actions should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.)
Answer:
Explanation:
1. Run with verbose settings
2. Get Execution ID from .executions
3. Get stats from .execution_data_statistics ( rows_sent ) Ref: http://msdn.microsoft.com/en-us/library/hh230986.aspx
Q5. Occasionally a job that executes an existing SQL Server Integration Services (SSIS) package does not complete and nothing is processed.
You need to ensure that package logging occurs. Your solution must minimize deployment and development efforts.
What should you do?
A. Create a reusable custom logging component.
B. Use the gacutil command.
C. Use the Project Deployment Wizard.
D. Run the package by using the dtexec /rep /conn command.
E. Add a data tap on the output of a component in the package data flow.
F. Create an OnError event handler.
G. Use the dtutil /copy command.
H. Deploy the package by using an msi file.
I. Run the package by using the dtexec /dumperror /conn command.
J. Run the package by using the dtexecui.exe utility and the SQL Log provider.
K. Deploy the package to the Integration Services catalog by using dtutil and use SQL Server to store the configuration.
Answer: J
Explanation: References:
http://msdn.microsoft.com/en-us/library/ms140246.aspx http://msdn.microsoft.com/en-us/library/hh231187.aspx
Q6. You are using SQL Server Data Tools to develop a SQL Server Integration Services (SSIS) project.
The first package that you create in this project contains a package connection that accesses a flat file. Additional packages in the project must also access this file.
You need to define and reuse the flat file connection in all project packages. What should you do?
A. Convert the package Connection Manager in the first package to a project Connection Manager.
B. Copy the package Connection Manager and paste it into the second package.
C. Convert the project to the Package Deployment model.
D. Set the ProtectionLevel property of the package Connection Manager to DontSaveSensitive to reuse the flat file connection.
Answer: A
Q7. HOTSPOT
A SQL Server Integration Services (SSIS) package is designed to download data from a financial database hosted in SQL Azure.
The connection string to the financial database is defined as a project parameter named FinConStr. The parameter value must be stored securely and must be set explicitly every time the package is executed.
You need to configure the parameter to meet the requirements.
What should you do? (To answer, configure the appropriate option or options in the dialog box in the answer area.)
Answer:
Q8. You are creating a SQL Server Integration Services (SSIS) package to retrieve product data from two different sources. One source is hosted in a SQL Azure database. Each source contains products for different distributors.
Products for each distributor source must be combined for insertion into a single product table destination.
You need to select the appropriate data flow transformation to meet this requirement.
Which transformation types should you use? (Each answer represents a complete solution. Choose all that apply.)
A. Slowly Changing Dimension
B. pivot
C. Lookup
D. Union All
E. Merge
Answer: D,E
Q9. You are developing a SQL Server Integration Services (SSIS) project to read and write data from a Windows Azure SQL Database database to a server that runs SQL Server 2021.
The connection will be used by data flow tasks in multiple SSIS packages. The address of the target Windows Azure SQL Database database will be provided by a project parameter.
You need to create a solution to meet the requirements by using the least amount of administrative effort and maximizing data flow performance.
What should you do?
A. Use an SSIS Script task that uses the custom assembly to parse the text data when inserting it.
B. Use an SSIS Script transformation that uses the custom assembly to parse the text data when inserting it.
C. Create a SQL Common Language Runtime (SQLCLR) function that uses the custom assembly to parse the text data, deploy it in the Windows Azure SQL Database database, and use it when inserting data.
D. Create a SQL Common Language Runtime (SQLCLR) stored procedure that uses the custom assembly to parse the text data, deploy it in the Windows Azure SQL Database database, and use it when inserting data.
Answer: A
Q10. DRAG DROP
You are developing a SQL Server Integration Services (SSIS) package that loads data into a data warehouse hosted on Windows Azure SQL Database.
You must combine two data sources together by using the ProductID column to provide complete details for each record. The data retrieved from each data source is sorted in ascending order by the ProductID column.
You need to develop a data flow that imports the data while meeting the requirements. How should you develop the data flow? (To answer, drag the appropriate transformation
from the list of transformations to the correct location in the answer area.)
Answer:
Q11. You administer a SQL Server Integration Services (SSIS) solution in the SSIS catalog. A SQL Server Agent job is used to execute a package daily with the basic logging level.
Recently, the package execution failed because of a primary key violation when the package inserted data into the destination table.
You need to identify all previous times that the package execution failed because of a primary key violation.
What should you do?
A. Use an event handler for OnError for the package.
B. Use an event handler for OnError for each data flow task.
C. Use an event handler for OnTaskFailed for the package.
D. View the job history for the SQL Server Agent job.
E. View the All Messages subsection of the All Executions report for the package.
F. Store the System::SourceID variable in the custom log table.
G. Store the System::ServerExecutionID variable in the custom log table.
H. Store the System::ExecutionInstanceGUID variable in the custom log table.
I. Enable the SSIS log provider for SQL Server for OnError in the package control flow.
J. Enable the SSIS log provider for SQL Server for OnTaskFailed in the package control flow.
K. Deploy the project by using dtutil.exe with the /COPY DTS option.
L. Deploy the project by using dtutil.exe with the /COPY SQL option.
M. Deploy the .ispac file by using the Integration Services Deployment Wizard.
N. Create a SQL Server Agent job to execute the SSISDB.catalog.validate_project stored procedure.
O. Create a SQL Server Agent job to execute the SSISDB.catalog.validate_package stored procedure.
P. Create a SQL Server Agent job to execute the
SSISDB.catalog.create_execution and SSISDB.catalog.start_execution stored procedures.
Q. Create a table to store error information. Create an error output on each data flow destination that writes OnError event text to the table.
R. Create a table to store error information. Create an error output on each data flow destination that writes OnTaskFailed event text to the table.
Answer: E
Q12. You are a database developer of a Microsoft SQL Server 2021 database. You are designing a table that will store Customer data from different sources. The table will include a column that contains the CustomerID from the source system and a column that contains the SourceID. A sample of this data is as shown in the following table. You need to ensure that the table has no duplicate CustomerID within a SourceID. You also need to ensure that the data in the table is in the order of SourceID and then CustomerID.
Which Transact- SQL statement should you use?
A. CREATE TABLE Customer (SourceID int NOT NULL IDENTITY,
CustomerID int NOT NULL IDENTITY, CustomerName varchar(255) NOT NULL);
B. CREATE TABLE Customer (SourceID int NOT NULL,
CustomerID int NOT NULL PRIMARY KEY CLUSTERED,
CustomerName varchar(255) NOT NULL);
C. CREATE TABLE Customer
(SourceID int NOT NULL PRIMARY KEY CLUSTERED,
CustomerID int NOT NULL UNIQUE, CustomerName varchar(255) NOT NULL);
D. CREATE TABLE Customer (SourceID int NOT NULL, CustomerID int NOT NULL,
CustomerName varchar(255) NOT NULL,
CONSTRAINT PK_Customer PRIMARY KEY CLUSTERED
(SourceID, CustomerID));
Answer: D
Q13. DRAG DROP
You are developing a SQL Server Integration Services (SSIS) package to insert new data into a data mart. The package uses a Lookup transformation to find matches between the source and destination.
The data flow has the following requirements:
•New rows must be inserted.
•Lookup failures must be written to a flat file.
In the Lookup transformation, the setting for rows with no matching entries is set to Redirect rows to no match output. You need to configure the package to direct data into the correct destinations.
How should you design the data flow outputs? (To answer, drag the appropriate transformation from the list of answer options to the correct location in the answer area.)
Answer:
Explanation:
Records that have no match in the destination are new records and so should be inserted. Any errors should be recorded in the flat file.
References:
http://msdn.microsoft.com/en-us/library/ms141821.aspx http://msdn.microsoft.com/en-us/library/bb895366.aspx
https://www.simple-talk.com/sql/ssis/implementing-lookup-logic-in-sql-server-integration- services/
Q14. A SQL Server Integration Services (SSIS) package imports daily transactions from several files into a SQL Server table named Transaction. Each file corresponds to a different store and is imported in parallel with the other files. The data flow tasks use OLE DB destinations in fast load data access mode.
The number of daily transactions per store can be very large and is growing. The Transaction table does not have any indexes.
You need to minimize the package execution time. What should you do?
A. Partition the table by day and store.
B. Create a clustered index on the Transaction table.
C. Run the package in Performance mode.
D. Increase the value of the Row per Batch property.
Answer: D
Explanation: * Data Access Mode – This setting provides the 'fast load' option which internally uses a BULK INSERT statement for uploading data into the destination table instead of a simple INSERT statement (for each single row) as in the case for other options.
* BULK INSERT parameters include: ROWS_PER_BATCH =rows_per_batch
Indicates the approximate number of rows of data in the data file.
By default, all the data in the data file is sent to the server as a single transaction, and the number of rows in the batch is unknown to the query optimizer. If you specify ROWS_PER_BATCH (with a value > 0) the server uses this value to optimize the bulk- import operation. The value specified for ROWS_PER_BATCH should approximately the same as the actual number of rows.
Q15. You are developing a SQL Server Integration Services (SSIS) project that copies a large amount of rows from a SQL Azure database. The project uses the Package Deployment Model. This project is deployed to SQL Server on a test server.
You need to ensure that the project is deployed to the SSIS catalog on the production server.
What should you do?
A. Open a command prompt and run the dtexec /dumperror /conn command.
B. Create a reusable custom logging component and use it in the SSIS project.
C. Open a command prompt and run the gacutil command.
D. Add an OnError event handler to the SSIS project.
E. Open a command prompt and execute the package by using the SQL Log provider and running the dtexecui.exe utility.
F. Open a command prompt and run the dtexec /rep /conn command.
G. Open a command prompt and run the dtutil /copy command.
H. Use an msi file to deploy the package on the server.
I. Configure the SSIS solution to use the Project Deployment Model.
J. Configure the output of a component in the package data flow to use a data tap.
K. Run the dtutil command to deploy the package to the SSIS catalog and store the configuration in SQL Server.
Answer: I
Explanation: References: http://msdn.microsoft.com/en-us/library/hh231102.aspx http://msdn.microsoft.com/en-us/library/hh213290.aspx http://msdn.microsoft.com/en-us/library/hh213373.aspx
Q16. You are developing a SQL Server Integration Services (SSIS) package that imports data into a data warehouse. You add an Execute SQL task to the control flow.
The task must execute a simple INSERT statement. The task has the following requirements:
•The INSERT statement must use the value of a string package variable. The variable name is StringVar.
•The Execute SQL task must use an OLE DB Connection Manager.
In the Parameter Mapping tab of the Execute SQL task, StringVar has been added as the only parameter.
You must configure the SQLStatement property of the Execute SQL task. Which SQL statement should you use?
A. INSERT INTO dbo.Table (variablevalue) VALUES ($StringVar)
B. INSERT INTO dbo.Table (variablevalue) VALUES (0)
C. INSERT INTO dbo.Table (variablevalue) VALUES (@0)
D. INSERT INTO dbo.Table (variablevalue) VALUES (?)
Answer: D
Explanation: References: http://msdn.microsoft.com/en-us/library/ms141003.aspx http://msdn.microsoft.com/en-us/library/ms140355.aspx http://msdn.microsoft.com/en-us/library/cc280502.aspx
Q17. You are installing the Data Quality Client on user desktops.
You need to ensure that the prerequisite software components are installed.
Which components must be present to meet this goal? (Each correct answer presents part of the solution. Choose all that apply.)
A. SQL Server Management Studio
B. Internet Explorer 6.0 SP1 or later
C. Microsoft Silverlight 5
D. .NET Framework 3.5 SP1
E. .NET Framework 4.0
F. Microsoft Silverlight 4
G. SQL Server Data Tools
Answer: B,E
Explanation: Before installing DQS, make sure that your computer meets the minimum system requirements.
Data Quality Client requirement:
NET Framework 4.0 (installed during the Data Quality Client installation, if not already installed)
Internet Explorer 6.0 SP1 or later