I'm working on a solution in Visual Studio that has 3 SSIS packages, and now I want to add a 4th that needs access to variables defined in one of the other packages. How can I do that?
hi, Is there any way to export a set of variables from each other package?? Issues comes when you have to built five packages which share the same variables. It's very tedious and hard to re-write them every time... Thanks for your time,
I have a simple task that I would like to manage through a SSIS package but do not know how to accomplish it. I need to perform the following tasks:
1. update a sql server table (ecwcust) and set the ftpstatus column to 'P' when the column's value is 'E'. update ecwcust set ftpstatus = 'P' where ftpstatus = 'E' 2. declare a variable and set the variable to the number of rows that have a ftpstatus equal to 'P' declare @newcustomercount int select @newcustomercount = count(*) from ecwcust where ftpstatus = 'P' 3. if @newcustomercount > 0, copy all of the rows from ecwcust where ftpstatus = 'P' to an excel spreadsheet. 4. send an email to an email recipient with the excel spreadsheet as an attachment. 5. delete the excel spreadsheet.
I do not know how to create the variable @newcustomercount so that I can determine whether or not to create the excel spreadsheet and copy the records to it. I have read all the online help and stuff that MS has but I can't understand the instructions. Please point me to some instructions that are understandable.
When executing a package from Business Intelligence Studio variables are well passed between packages but when same is done using dtexec utility variables don't get passed between packages. Does anyone have any idea why this could be happening and what would be possible solutions? Thanks in advance!
We're just in the middle of performing our first release of SSIS packages through various environments.
The way we are set up currently is the developer will check the package(s) and related config files out of source control, develop on their own machine and check everything in again. Then we deploy the packages consecutively to the Dev, Tst and Prd servers.
We are going down the path of using one environment variable for every config file. some packages share config files (e.g. we've only one config file for each database or ftp connection etc.) and some config files are package specific (error log file connections and success/failure e-mail sources etc.).
What we want ideally is a script that we can check into source control that will create the environment variables on a server at deployment. The "set" command at the command line can be used to change the value of an environment variable, or to create a session-specific variable, but not to create environment variables.
So far the only method that we're using is manually typing in the environment variable names via control panel and copying and pasting the paths into the value fields. Given that we're deploying potentially hundreds of config files, it's obvious that this new-fangled GUI point-and-click and copy-and-paste method of deployment is absolutely foolproof and totaly not prone to any error whatsoever.
Please tell me there's a way to create and set environment variables without going through control panel. running a script or something to do it automatically will: Ensure that each environment is set up accurately and identically, eliminating human error. Ensure that when a developer checks out a package to their local drive, although they may have to change the variable values, he can at least create the relevant variables without having to type them in. Enable efficient migration to another new server (for example during Disaster Recovery).
Can anyone point me to some example scripts at all?
It is pretty easy to create the package configuration to the shared, never-changing path to XML file. But to get this to work, I have to add a bunch of package variables which are driven from the configuration. I then use expressions to compute actual paths to source files and packages.
This business of copying a block of variables from one package to another is extremely tedious, and at least to me, something of a maintenance issue. I have found if I get ahead of myself and forget to create the variables first, then Visual Studio is NOT very forgiving, and will complain repeatedly just prior to crashing without the offer to save anything. If I add a new variable, it is like an easter egg hunt. This does NOT feel very single point of maintenance when variables must be used in this way to really effectively apply configuration, and it is so tedious and error prone, and unforgiving. I seem to spend a great percentage of time copying the names and datatypes and values of variables ONE AT A TIME. What is that all about?
I hear people telling the praises of configurations, but I must be missing something.
With DTS I used a system where global scope settings were "injected" into packages as they ran, and I never had to worry about anything, and it was damn hard to goof it.
Please someone, evangelize me!! Are people bypassing VS and using NOTEPAD on the packages or something like that to save time?
I am not comfortable with DTS 2000 but I need to execute a encapsulated DTS 2000 package from a SSIS package. The real problem is when I need to pass SSIS variables to DTS 2000 package. The DTS 2000 package have 3 global variables that I can identify on " Execute DTS 2000 Package Task Editor - Inner Variables ". I believe the SSIS variables must be mapped on " Execute DTS 2000 Package Task Editor - OuterVariables ". How can I associate the SSIS variables(OuterVariables ) to "Inner Variables"? How can I do it? Much Thanks.
I've run into a problem with SSIS packages wherein tasks that write or copy files, or create or delete directories, quit execution without any hint of an error nor a failure message, when called from an ASP.NET 2.0 application running on any other machine than the one where the package was created from. By all indications it appeared to be an identity/permissions problem.
Our application involves a separate web server and database server. Both have SQL Server 2005 installed, but the application server originally only had Integration services. The packages are file system-deployed on the application server, and are called using Microsoft.SqlServer.Dts.Runtime methods. For all packages that involve file system tasks, the above problem occurs.
When the above packages are run using the command prompt (either DTEXEC or DTEXECUI) the packages execute just fine. This is expected since we are using an administrative account. However when a ShellExecute of the same command is called from ASP.NET, the same problem occurs.
I've tried giving administrative permissions to the ASPNET worker process user to no avail.
I have likewise attempted to use the SQL Server Agent job approach but that approach might not be acceptable for our clients since it means installing SQL Server 2005 Database services on the application server.
I have read the relevant threads in this forum, namely http://forums.microsoft.com/MSDN/ShowPost.aspx?PostID=1044739&SiteID=1 and http://forums.microsoft.com/MSDN/ShowPost.aspx?PostID=927084&SiteID=1 but failed to find any solution appropriate for our set up.
We manage some SSIS servers, which has only SSIS and SSIS tools installed on them and not the sql server DB.
SSIS packages and configuration files are deployed on a NAS. We run the SSIS packages through DTEXEC by logging in to the server.
We want to allow developers to run their packages on their own on the server, but at the same time we dont want to give them physical access on the server i.e we do not want to add them into RDP users list on server properties. We want them to allow running their packages remotely on the server.
One way We could think of is by using powershell remoting and we are working on that. But is there any other way or any tool already present for the same.
Environment: Windows Server 2003. 2 node Cluster single instance of SQL Server 2K sp3a. 1 SAN.
Cluster groups are: - Cluster group with its own default resources. - Oracle group with its own resources. - SQL Server group with its own resources (Server,services,disks,ip address....) Disk v: is for SQL Server databases Disk w: is for SQL Server logs Disk x: is for SQL Server backups.
SO now I have to create a shared folder where the clients need to connect in order to launch the program (executable file *.exe) and maybe tomorrow I will also need to create others shared folders where to put others executable. Questions: 1)Is it correct to create theses shared folders in the SQL Server group? 2) In wich one of the tree disks? 3) If I have 20 differents programs with differents security will I end up in having a SQL Server group with 20 different resources (file shares)? Please advise. Kind regards. Franco
Is anyone using sql server 70 or 2000 and a EMC or LSI Metastore SAN? If so, I would like to know how things are going for you. We are currently evaluating both of these products and will be making a decision as to which we will purchase in February. Please advise if you have anything positive or negative to offer. Thanks. Gail Wade gwade@it.rjf.com
I have setup a Microsoft SQL 7.0 cluster as Actice/Passive configuration on NT 4.0 Enterprise. The problem I am having is that when I create a shared folder on Node B on a shared drive then I reboot Nobe B, the cluster fails over to Node A, but the shared dispeares.
Hopefully a painless and quick question to answer:
Am backing up the SQL2000 database(s) using the maintenance plan's under management in the Enterprise Manager. All is fine however, we are running out of diskspace as the database grows, which is causing the backup to fail if I don't manually remove the nights before.
I have selected to delete old backup files however these files are deleted after the new file is generated and due to the lack of space the new one is not being generated....so in effect the older ones never get a chance to be deleted.
So I was wanting to know if it is at all possible to backup to a UNC path or share . I have tried however it raises an error saying it is not a valid path.
Any assistance here is hugly appreciated (asides from getting new disk - as this isn't a short term option at the moment)...
Can someone please tell me what is meant by a file share?I'm doing a little reading on MS Sql Server and they talk about howyou can connect to the sql server through a named pipe.I'm guessing that a file share means a specific service running on aserver. Is this correct?The book says the following:When you connect to a file share you use a Universal Naming Conventionpath to a file server: \FILESERVERSharename.To connect to a named pipe, you connect to a share of the form:\COMPUTERpipesqlqueryThanks Hope you can clear this up for me.
w3wp!library!1!01/29/2007-17:15:39:: e ERROR: Throwing Microsoft.ReportingServices.Diagnostics.Utilities.ServerConfigurationErrorException: The report server has encountered a configuration error. See the report server log files for more information., SharePoint content service is null. Report Server may not have joined the SharePoint farm, or Report Server service account may not have been granted access to farm.; Info: Microsoft.ReportingServices.Diagnostics.Utilities.ServerConfigurationErrorException: The report server has encountered a configuration error. See the report server log files for more information.
I tried to make the report server integrate with share point 2007. all services account are the domain account and local admin, including the applicaiton pool in IIS.
i'm trying to write a script just like "copy database wizard" operation, for detaching, copy and ataching a database from a remote sql server to a local server.
Now the problem is when i try to copy the mdf file from the sql remote server to my local server.
Is it possible to create a UNC share in the remote sql server and delete after with t-sql like the Copy Database wizard does?
I would like to backup my databases to a network share (NAS) instead of local disk using Maintenance Plans created by Enterprise Manager. I have successfully used a UNC path to target the destination network share but have not been to figure out how to submit a logon to the network share before the backup is executed.
The SQL Server instance is running in the context of the local system account.
Can I insert a step in the SQL SQL job that is created by the Maintenance Plans that changes the Windows account that the backup runs under? If yes what command syntax would I use in the inserted step or is there another way to accomplish that I'm attempting to do?
I am trying to backup a database with a command like:
BACKUP DATABASE my DataBase TO DISK = '\bkSystemkDiskBackup1.bak'
but I get the error 'Cannot open backup device '\bkSystemkDiskBackup1.bak'. Device error or device off-line. The bkDisk folder is shared, with Everyone full-control access (it's a test environment)
I have SQL Server 2000 database involving voluminous data spread across different locations in a state. The database involves lot of transactions. The database involves tables storing photo images and thumb impressions along with the textual data and other needed information in a single table. I have a application accessing and updating about 5 tables and reading most of the master tables. I need a strategy to design the database such that performance improves. Once the photo and thumb are inserted for a particular party it is not required for retrieval immediately at any circumstance. In this scenario how should I design the database and keep data volume in check. Can I archive photo and thumb impressions in CDs. Will it be reliable and feasible. Please clarify on this issue. Please share your valuable experience. Only insertions involved in the partyinfo table having Photo and Thumb impressions. Later the table is only used to fetch the data of party other than Photo and thumb.
Hi,I am not able to execute following command from computers other than the SQLserver itself.select * from openrowset(bulk N'\ImageServerModelCam_05.dwg',SINGLE_BLOB) as docI get following error message:Msg 4861, Level 16, State 1, Line 1Cannot bulk load because the file "\ImageServerModelCam_05.dwg " couldnot be opened. Operating system error code 5(Access is denied.).If I use local file path and keep the file on the SQL server then it worksfine from any computer.select * from openrowset(bulk N'D:ModelCam_05.dwg', SINGLE_BLOB) as docThe network share has read access for everyone.What should I do to make it work? I am not allowed to move the files and theapplication which loads them to the SQL server.ThanksSachin G
I have got a project where I am asked to utilize SharePoint features as much as possible. This project is earlier developed in .Net using Sharepoint Object Model. But now .Net should be eliminated as much as possible.
The requirements are : Display a list of projects for the logged in user (windows logged in user). The list should have search criteria: project name, customer, project manager, project period.
On click of a project in the list, project details should be displayed Details like customer, team, start date, highlights, news. On click of a team member's name, Details of the team member (Name, office location, project joining, releaving dates, photo ) should be displayed.
To do this in Sharepoint my major hurdle is to display project list for windows logged in user. How do I do this in Sharepoint???
Expecting ideas/workarounds from all... Please be kind enough for that.
I'm new to replication. Can anyone help with the following?
I want to publish the specific rows of data that are created on the publisher without accepting changes to those rows from subscribers, but do accept new rows (and changes to those rows) from subscribers.
EG. publisher creates rows 1 and 2 which are published to the subscriber
subscriber creates rows 3 and 4 which are merged back to the publisher
subscriber's updates (or delets) to rows 1 and 2 are not merged back to the publisher
So basically I want the publisher created rows to be published but remain as is whilst allowing subscriber created rows to be merged back to the publisher.
I've recently moved an asp.net website from my PC to a network share because another tech it going to be working on it. I finally got the correct permissions on the network share and the correct .NET Framework settings on my PC to be able to run the app. Now I can't access the SQL server which is on a different server. Getting the following error: Request for the permission of type 'System.Data.SqlClient.SqlClientPermission, System.Data, Version=2.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089' failed. How to I setup access to my SQL server for the app from any given PC on my LAN?
I make code like this: In Application_Start,Create a SqlConnection and store it into HttpApplication:HttpContext.Current.Application["SQLCONN"] = new SqlConnection("XXXXConnectionString"); ((SqlConnection)HttpContext.Current.Application["SQLCONN"]).Open();
While using this connection , use it as Sqlconnection(what is already opened), and do not close it: SqlCommand db_comm = new SqlCommand("dbo.abcd",(SqlConnection)HttpContext.Current.Application["SQLCONN"]); After Application_End,Close this Sqlconnection: ((SqlConnection)HttpContext.Current.Application["SQLCONN"]).Close(); ========================== This can save a lot of trouble to manage connections,i think. but i don't know whether it will cause connection pool errors.
how can Share may sqlexpress 2005 on network that other can access my databasess on their pc and edit update and insert recored into my table please tell me how can do this ?
I'm trying to run a two-computer upgrade from 6.5 to 7.0. In the Upgrade Wizard I receive the message "The Wizard was unable to connect to the ADMIN$ share on your remote 6.x SQL Server. Please verify that this share exists and that you have sufficient permissions to access it." The share does exist and the login should have sufficient permissions. To which login is this message referring? Is it the NT domain login on the import machine or is it one of the services on the export machine?