Master the 70 463 training kit pdf Implementing a Data Warehouse with Microsoft SQL Server 2012 content and be ready for exam day success quickly with this Examcollection 70 463 training kit pdf real exam. We guarantee it!We make it a reality and give you real 70 463 exam questions in our Microsoft 70 463 training kit pdf braindumps.Latest 100% VALID Microsoft 70 463 pdf Exam Questions Dumps at below page. You can use our Microsoft microsoft 70 463 braindumps and pass your exam.
♥♥ 2021 NEW RECOMMEND ♥♥
Free VCE & PDF File for Microsoft 70-463 Real Exam (Full Version!)
★ Pass on Your First TRY ★ 100% Money Back Guarantee ★ Realistic Practice Exam Questions
Free Instant Download NEW 70-463 Exam Dumps (PDF & VCE):
Available on:
http://www.surepassexam.com/70-463-exam-dumps.html
Q1. You are designing a SQL Server Integration Services (SSIS) package. The package moves order-related data to a staging table named Order. Every night the staging data is truncated and then all the recent orders from the online store database are inserted into the staging table.
Your package must meet the following requirements:
. If the truncate operation fails, the package execution must stop and report an error. . If the Data Flow task that moves the data to the staging table fails, the entire refresh operation must be rolled back. . For auditing purposes, a log entry must be entered in a SQL log table after each execution of the Data Flow task.
The TransactionOption property for the package is set to Required.
You need to design the package to meet the requirements.
How should you design the control flow for the package? (To answer, drag the appropriate setting from the list of settings to the correct location or locations in the answer area.)
Answer:
Q2. You are designing a SQL Server Integration Services (SSIS) data flow to load sales transactions from a source system into a data warehouse hosted on Windows Azure SQL Database. One of the columns in the data source is named ProductCode.
Some of the data to be loaded will reference products that need special processing logic in the data flow.
You need to enable separate processing streams for a subset of rows based on the source product code.
Which Data Flow transformation should you use?
A. Multicast
B. Conditional Split
C. Script Task
D. Data Conversion
Answer: C
Q3. You are developing a data flow transformation to merge two data sources. One source contains product data and the other source contains data about the country in which the product was manufactured. Both data sources contain a two-character CountryCode column and both use SQL Server. Both data sources contain an ORDER BY clause to sort the data by the CountryCode column in ascending order.
You use a Merge Join transformation to join the data.
You need to ensure that the Merge Join transformation works correctly without additional transformations.
What should you do? (Each correct answer presents part of the solution. Choose all that apply.)
A. Change the ORDER BY clause on the product source to order by ProductName.
B. Change the Merge Join transformation to a Merge transformation.
C. Set the appropriate SortKeyPosition properties on the data sources.
D. Set the IsSorted property on both data sources.
Answer: C,D
Q4. You are developing a SQL Server Integration Services (SSIS) package that imports data into a data warehouse hosted on SQL Azure.
The package uses a Foreach container to process text files found in a folder. The package must be deployed to a single server by using the Project Deployment model.
Multiple SQL Server Agent jobs call the package. Each job is executed on a different schedule.
Each job passes a different folder path to the package.
You need to configure the package to accept the folder path from each job.
Which package configuration should you use?
A. Parent Package Variable
B. XML Configuration File
C. Environment Variable
D. .dtsConfig file
E. Registry Entry
Answer: C
Q5. You are troubleshooting an existing SQL Server Integration Services (SSIS) package. On several occasions, the package execution does not finish and no data seems to have been transferred.
You need to ensure that package logging occurs. Your solution must minimize deployment and development efforts.
What should you do?
A. Run the dtutil command to deploy the package to the SSIS catalog and store the configuration in SQL Server.
B. Use an msi file to deploy the package on the server.
C. Open a command prompt and execute the package by using the SQL Log provider and running the dtexecui.exe uti
D. Create a reusable custom logging component and use it in the SSIS project.
E. Open a command prompt and run the dtutil /copy command.
F. Configure the SSIS solution to use the Project Deployment Model.
G. Add an OnError event handler to the SSIS project.
H. Open a command prompt and run the gacutil command.
I. Open a command prompt and run the dtexec /dumperror /conn command.
J. Open a command prompt and run the dtexec /rep /conn command.
K. Configure the output of a component in the package data flow to use a data tap.
Answer: G
Explanation:
References:
http://msdn.microsoft.com/en-us/library/ms141212.aspx
http://www.mssqltips.com/sqlservertip/2450/ssis-package-deployment-model-in-sql-server-2012-part-1-of-2/
http://www.mssqltips.com/sqlservertip/2450/ssis-package-deployment-model-in-sql-server-2012-part-2-of-2/
Q6. You are implementing a new SQL Server Integration Services (SSIS) 2012 package that loads data from various flat files and a Windows Azure SQL Database database.
Daily transactions must be loaded into a staging database. All the SSIS tasks will use the CurrentDate variable as the transaction date.
You need to set the CurrentDate variable to the current execution date when the package starts. You need to achieve this goal by using the least amount of development effort.
What should you use to set the variable?
A. An Expression task
B. A Script component
C. A Script task
D. An Execute SQL task
Answer: A
Q7. You are completing the installation of the Data Quality Server component of SQL Server Data Quality Services (DQS).
You need to complete the post-installation configuration.
What should you do?
A. Install the Analysis Services OLE DB Provider.
B. Make the data available for DQS operations.
C. Run the Data Quality Server Installer.
D. Install ADOMD.NET.
Answer: C
Explanation:
References: http://msdn.microsoft.com/en-us/library/ff877917.aspx
http://msdn.microsoft.com/en-us/library/gg492277.aspx
Q8. You are developing a project that contains multiple SQL Server Integration Services (SSIS) packages. The packages will be deployed to the SSIS catalog. One of the steps in each package accesses an FTP site to download data files.
You create project parameters to store the username and password that are used to access the FTP site.
You need to ensure that the username and password values are encrypted when they are deployed.
What should you do?
A. Convert the parameters to package parameters.
B. Set the Sensitive property of the parameters to True.
C. Set the ProtectionLevel property of the package to EncryptSensitiveWithPassword.
D. Convert the project to the Legacy Deployment model.
Answer: B
Q9. You administer a Microsoft SQL Server database. You want to import data from a text file to the database.
You need to ensure that the following requirements are met:
. Data import is performed by using a stored procedure.
. Data is loaded as a unit and is minimally logged.
Which data import command and recovery model should you choose? (To answer, drag the appropriate data import command or recovery model to the appropriate location or locations in the answer area. Each data import command or recovery model may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.)
Answer:
Q10. You are developing a SQL Server Integration Services (SSIS) package to load data into a data warehouse. The package consists of several data flow tasks.
The package experiences intermittent errors in the data flow tasks.
If any data flow task fails, all package error information must be captured and written to a SQL Server table by using an OLE DB connection manager.
You need to ensure that the package error information is captured and written to the table.
What should you do?
A. Use an event handler for OnError for the package.
B. Use an event handler for OnError for each data flow task.
C. Use an event handler for OnTaskFailed for the package.
D. View the job history for the SQL Server Agent job.
E. View the All Messages subsection of the All Executions report for the package.
F. Store the System::SourceID variable in the custom log table.
G. Store the System::ServerExecutionID variable in the custom log table.
H. Store the System::ExecutionInstanceGUID variable in the custom log table.
I. Enable the SSIS log provider for SQL Server for OnError in the package control flow.
J. Enable the SSIS log provider for SQL Server for OnTaskFailed in the package control flow.
K. Deploy the project by using dtutil.exe with the /COPY DTS option.
L. Deploy the project by using dtutil.exe with the /COPY SQL option.
M. Deploy the .ispac file by using the Integration Services Deployment Wizard.
N. Create a SQL Server Agent job to execute the SSISDB.catalog.validate_project stored procedure.
O. Create a SQL Server Agent job to execute the SSISDB.catalog.validate_package stored procedure.
P. Create a SQL Server Agent job to execute the
SSISDB.catalog.create_execution and SSISDB.catalog.start_execution stored procedures.
Q. Create a table to store error information. Create an error output on each data flow
destination that writes OnError event text to the table.
R. Create a table to store error information. Create an error output on each data flow
destination that writes OnTaskFailed event text to the table.
Answer: I
