You have a SQL Server Integration Services (SSIS) package. The package contains a script task that has the following comment.
// Update DataLoadBeginDate variable to the beginning of yesterday
The script has the following code.
Dts.variables[''User::DataLoadBeginDate''].Value = DataTime.Today.AddDays(-1);
The script task is configured as shown in the exhibit. (Click the Exhibit button.)
When you attempt to execute the package, the package fails and returns the following error message: ''Error: Exception has been thrown by the target of an invocation.''
You need to execute the package successfully.
What should you do?
You add existing variables to the ReadOnlyVariables and ReadWriteVariables lists in the Script Task Editor to make them available to the custom script. Within the script, you access variables of both types through the Variables property of the Dts object.
References:
https://docs.microsoft.com/en-us/sql/integration-services/extending-packages-scripting/task/using-variables-in-the-script-task?view=sql-server-2017
You are developing a SQL Server Integration Services (SSIS) package.
To process complex scientific data originating from a SQL Azure database, a custom task component is added to the project.
You need to ensure that the custom component is deployed on a test environment correctly.
What should you do?
You are developing a SQL Server Integration Services (SSIS) project that copies a large amount of rows from a SQL Azure database. The project uses the Package Deployment Model. This project is deployed to SQL Server on a test server.
You need to ensure that the project is deployed to the SSIS catalog on the production server.
What should you do?
To facilitate the troubleshooting of SQL Server Integration Services (SSIS) packages, a logging methodology is put in place.
The methodology has the following requirements:
* The deployment process must be simplified.
* All the logs must be centralized in SQL Server.
* Log data must be available via reports or T-SQL.
* Log archival must be automated.
You need to configure a logging methodology that meets the requirements while minimizing the amount of deployment and development effort.
What should you do?
You are designing an enterprise star schema that will consolidate data from three independent data marts. One of the data marts is hosted on SQL Azure.
Most of the dimensions have the same structure and content. However, the geography dimension is slightly different in each data mart.
You need to design a consolidated dimensional structure that will be easy to maintain while ensuring that all dimensional data from the three original solutions is represented.
What should you do?
Currently there are no comments in this discussion, be the first to comment!