Setup And Upgrade :: Several Same Environment Variables In Path
May 5, 2015
I noticed accidentally today on my PATH and discovered that I have several variables for different sql versions. I can understand this as I have upgraded from 2008, to 2012 then 2014. Here are the list of variables
If you have your Data, Logs, System and TempDB all in VMDK's and those VMDK's are formatted to 64K and then reside on Clustered Storage that is formatted as 4K, which is then running through a SAN controller that is reading and writing in 2MB chunks, is there value in formatting the SQL drives as 64K? Also, would it be better to format the Clustered Storage as 64K? Is this a situation in which formatting the SQL drives as 64K is becoming a moot point?
We're doing upgrades from SQL 2008 R2 to SQL 2014. This is blocked due to RS is installed but not configured. Our desired action is to uninstall RS and proceed with the upgrade. But when setuparp.exe is raised, it does not list all the features on the 'Select Features' page. In fact, it only lists the last 2 shared features (SQ Client Connectivity SDK and Microsoft Sync Framework). However, all items appear to be listed on the 'Select Instance' page including RS. I've seen this issue on 2 of our SQL 2008 R2 Servers already.
I'm trying to install SQL Server Management studio 2012 on my Windows 7 (x64) standalone laptop. When I click "New SQL stand-alone installation..." it runs a Setup Support Rules check and always fails "Setup Account Privileges". I've looked into the error and I keep getting that I need to change security rules but I don't have that option in window 7. How do I get around this without having to resort to a computer running Windows Server?
I have Visual Studio 2013 premium installed along with Localdb v11. I just want to connect and manage my database engine through SSMS when developing any application.
I have about 40 DTS packages that I want to run against three different databases on the same server. Can someone suggest an easy way to run these jobs and differentiate each time which SQL or INI file to use.
I am running a number of SQL instances on my PC. Within the network, I have think server with various System Center components. For compatibility reasons, some features of System Center 2012 R2 had to be delegated to different SQL databases. My question is, because there is now more than one IP address on my system, and each instance of SQL is assigned to its own IP, is there a way to setup DNS and SQL so the namespace points to the desired IP address? For Instance:
MSSQL2008 instance is set to run on = 11.12.13.1 MSSQL2012 is set to run on = 11.12.13.2 IN DNS: A Record: Mike-PC = 11.12.13.1 A Record: Mike-PC = 11.12.13.2
If I want to use MSSQL2008 by specifying Mike-PC as the DNS name, how would I do that with 100% accuracy? If there is another way to get the job done, I am more than willing to approach this differently.
I'm currently using the setup below for my RDA testing. Is there any problem with this setup?
connected connected Client Apps (SQLCE 3.0) -----------------> Windows Server 2003 (Web server) ----------------> Windows XP (SQL2005)
previously was using windows XP to act as web server and have SQL server reside inside the same machine but having 10 connection limitation from windows xp where only 10 client can connect to the machine. The current setup moves the web server out to windows server 2003 and i believe it solves the 10 concurrent connection limitation but how about the SQL server? is it limited to 10 connection from windows server 2003 also? (Windows XP is my testing server, Windows Server 2003 contains important test data for the whole company) . I tried with concurrent RDA pull for 40 client concurrently with 1 pc having 10 client apps but obtain not really satisfying results. Pulling one table (20Mb) needs 1 to 2 mins through LAN networks but pulling concurrently with 40 clients takes up to 20 - 30 mins.
does anybody know if it is possible to use environment variables when calling dtexec utility?
I'd like to run packages stored on server's file system from directory that I've had specified in an environment variable called SSIS_PackagesPath.
Now, I'am trying to write dtexec command, where path to the actual SSIS package would be concatenation of environment variable (i. e. path to package directory) and name of package itself (written explicitly). Is this syntactically possible?
The reason behind is to be able to easily modify package storage directory for multiple scheduled jobs that run SSIS packages.
I am looking for advice with setting up a 2005 environment for development, debugging and testing using a laptop as the development platform and a desktop pc as the database engine both connected via a LAN. The current configuration is that all the full blown Visual Studio / SQL development software is installed on the development laptop and can go wherever the developer goes, and the Express versions are installed on the desktop pc so that the development database has an unbroken connection with the network. This setup worked well in the past for developing applications that use a typical database interface. The applications could be completely developed and tested before it went live on the corporation€™s full blown production SQL server. Now we want to utilize the Reporting Services/SSIS in our development efforts and I would like to know what parts of the development software (Express vs. Full) should be installed on what unit (laptop vs. desktop) to fully develop and test using Reporting Services before we move to production? And anything else, like what services need to be running on what units etc.
I am migrating the BE of an Access app. to SQL server 2012. I need to get the user's login name (Windows Authentication login). This can be done using xp_cmdshell, but, xp_cmdshell is considered dangerous and I wouldn't be able to run it once I deploy the app. to the company servers (currently I have SQL server on my computer and as an admin I can enable xp_cmdshell to run, but IT doesn't allow it in company servers for security reasons).
Another question, is it possible to send data from the logged in user from Access to SQL server? What I need to do is let SQL know the username of the logged in user, then, use it to filter the data on SQL. Idea is that user can only run queries for his data (he can't view other user's data unless he is a manager or an admin (currently the app. in Access logs the user in automatically if his Windows Domain username is found in the user's table, and set's his role found in the Roles table). It is this functionality that is giving me some problems to migrate to SQL.
I created a function that uses the System_User SQL built-in function, this retrieves the SQL login username, but, the app. uses 1 SQL local account to connect to the server, so in essence it doesn't work as I need the Windows Domain account username.
We're just in the middle of performing our first release of SSIS packages through various environments.
The way we are set up currently is the developer will check the package(s) and related config files out of source control, develop on their own machine and check everything in again. Then we deploy the packages consecutively to the Dev, Tst and Prd servers.
We are going down the path of using one environment variable for every config file. some packages share config files (e.g. we've only one config file for each database or ftp connection etc.) and some config files are package specific (error log file connections and success/failure e-mail sources etc.).
What we want ideally is a script that we can check into source control that will create the environment variables on a server at deployment. The "set" command at the command line can be used to change the value of an environment variable, or to create a session-specific variable, but not to create environment variables.
So far the only method that we're using is manually typing in the environment variable names via control panel and copying and pasting the paths into the value fields. Given that we're deploying potentially hundreds of config files, it's obvious that this new-fangled GUI point-and-click and copy-and-paste method of deployment is absolutely foolproof and totaly not prone to any error whatsoever.
Please tell me there's a way to create and set environment variables without going through control panel. running a script or something to do it automatically will: Ensure that each environment is set up accurately and identically, eliminating human error. Ensure that when a developer checks out a package to their local drive, although they may have to change the variable values, he can at least create the relevant variables without having to type them in. Enable efficient migration to another new server (for example during Disaster Recovery).
Can anyone point me to some example scripts at all?
I have deployed a project with multiple packages to SSIS 2012 db. I am able to configure the project parameters fine. But, I am not able to replace the package variable values with the 'Environment' variables.
Any fix for the seemingly random sort order of the variables in the dropdown list when configuring parameters and connection managers in the SSISDB catalog?
I imported all of our connection strings into an environment (about 200 of them). They were inserted in alpha order and the ID values within the internal.environment_variables table shows them in order as well, by ID and by name. When I run profiler and capture the command that retrieves them and run it in ssms they are in order but in the dropdown they seem random.
There are no values within any of the tables that accounts for the order they are in.
If a package has 5 connections you need to go through the unsorted list 5 times to find them.
Sometimes you get lucky and they are in the first 20 or so.
I know I can write a script, just wondering if there is a fix for the sorting.
I have SSIS 2012 Enterprise, using catalog deployment and have more that 50 environment variables for connection to databases across my enterprise.
The problem when i go to configure the packages after deployment and pick the proper env variables, that are not sorted, so i have to browse all entries in order to find the proper entry in environment variables.
I am trying to set up log shipping in a clustered server environment. I am pretty confused about the location of the shared folder to be created to put backup created by  log shipping job. Which drive should I use either local or clustered shared drive to store the backups in primary server and  to copy the same in secondary Server?
We installed the SQL Server 2012 GUI tools, including "Upgrade Advisor," on a new virtual machine. We can connect to the old SQL Server 2008 R2 clustered databases just fine over port 1433 using SQL Server Management Studio 2012, but Upgrade Advisor is a mixed bag. Here are the steps we are taking:
- Click on "Start," "All Programs," "Microsoft SQL Server 2012," "SQL Server 2012 Upgrade Advisor"
- Fill in the server name with a valid SQL Server Network Name
- Select "SQL Server" as the only component to be analyzed
- Fill in the instance name with a valid named instance
- Use Windows Authentication (the default)
- The "SQL Server Parameters" window displays valid database names, meaning that we have a successful connection to the desired SQL Server instance at this point
- Leave all of the databased selected (the default)
- Click the "Run" button
- The "Upgrade Advisor Progress" window spins for a while and eventually generates the following error:
TITLE: mscorlib ------------------------------ The network path was not found. ------------------------------ BUTTONS: OK ------------------------------
Why would Upgrade Advisor be able to display the database names after making a successful connection only to fail later saying that it cannot find the network path?
We have installation of Dbase Engine and SSIS that is PRODUCTION, and want to replace with newer hardware. In "the old days", we built "boxname_new" and installed SQL with "sqlname_new", took PROD users off-line, and quickly renamed original boxes/SQL and new boxes/SQL to original name, copied data and off we went with upgrade.
NOW, the "renaming" option for SQL tools is not supported, but with re-installation.
Has anyone developed game plan steps for accomplishing hardware upgrade, including SQL environment swap with MINIMAL downtime for PRODUCTION environment? Can you share?
I got this error message during SQL Server Express install with no error code on Server 2003 SP1. I'm installing on an older development machine and although I do have a very large log file somewhere I wanted to post before I even dug into it. I noticed in IIS under my Webs that the MSSQLExpress webs were there at first but when I clicked on them to get any info or Properties... they dissapeared. So they were almost installed but they no longer exist. Also I do not have a SQL Service Manager icon down near the clock.
I almost took a step to fix this. After reading a few posts about similar problems I thought maybe this happened because the Default Web Sites IP # was not set to (All Unassigned). I was going to try a repair install from Add/Remove Programs but the temporary setup files are already gone. I really dont want to uninstall all 5 SQL components from Add/Remove Programs and start from scratch.
As good security practice we usually remove the Default Web or as in this development servers case we changed the IP from (All Unassigned) to an IP that is only accessible from certain machines on the internal network. I obviously should have read more pre setup guides. Anyway how can I fix this without reinstalling everything?
I am getting the error: "Cannot connect to WMI provider. You do not have permission or the server is unreachable."I tried the solution posted on the Microsoft support page involving a command prompt and the mofcomp "%programfiles(x86)%Microsoft SQL Server<var class="sbody-var"> numbe r</var>Sharedsqlmgmproviderxpsp2up.mof"...This still does not fix the problem, the result of running the above is: Parsing MOF file: C:Program Files (x86)Microsoft SQL Server110Sharedsqlmgmp roviderxpsp2up.mof...MOF file has been successfully parsed...Storing data in the repository..An error occurred while processing item 1 defined on lines 4 - 7 in file C:Prog
Is there way to completely re-install the Configuration Manager? One post suggested the Access denied was Administrator rights but I have full Administrative rights to everything so I don't see that being the problem.
So, I just downloaded and installed Windows 8 and SQL Server 2014 Developer. I used to have Visual Studio, that seems to be gone now. Can't find it. I can't figure out a way to get to SSIS or SSAS.
when i install sql server 2005 in cluster. I have error at Setup process screen --> the selected components are being configured.all the components are failed.I am using domain account with administrator account privilege on both server.OS : windows server 2008 on virtual machine..when i install sql server agent is enable to select during setup..
I am using windows 8 and I have only one user in laptop. While installing SQL 2014, I am getting an error (Missing System Administrator). Under SQL Server Agent and SQL Server Database I have select the system user and gave the password but still I am getting the error.
I'm trying to do a SQL 2008 cluster installation.I installed one node and now I'm trying to add a failover cluster node.in the "add Node Rules" I get following message:
Rule Check Result...Rule "SQL Server Database Services feature state" failed.The SQL Server Database Services feature failed when it was initially installed. The feature must be removed before the current scenario can proceed.
If I close SSMS, it restarts itself as if it crashed and trying to restore itself. Even if I hit cancel as it's restarting, it will restart itself yet again. This happens twice. what's causing this?