RollBack Analysis Processing Task

Mar 7, 2008



Hi there,

i have got a SSIS Package, that contains a sequence container with transactionoption "required". Within this sequence I placed different AS processing tasks and different SQL tasks. The transactionoptions of these tasks are set to supported.

My problem: in the case a SQL task fails on execution, all executed tasks are rolled back except the AS processing tasks. The expected and necessary behavior should be, that also the AS processing tasks get rolled back.

Has anyone got a solution or a workaround for this problem?

Thanks.
Andi

View 3 Replies


ADVERTISEMENT

Analysis :: Is There Any Rollback Operation Whenever Cube Processing Fail

Apr 24, 2014

I Have one cube, First time its Processed successfully and browsed fine. Now i was change in the cube Calculations/somewhere, now again process the process failed. I am unable to browse my cube. Is there any option to Rollback this processed cube and Browse with existing cube?What is require is, If cube process is fail, i want to browse the cube with existing cube after modify the cube also(if fails).

View 9 Replies View Related

Analysis :: Cube Needs To Be Deployed From VS After SSIS Analysis Services Processing Task Completes?

May 13, 2014

I have a cube that we are processing nightly via an Analysis Service Processing Task in SSIS.  In order to increase the performance of the processing time, we elected to use a lot of rigid dimension attributes, and do a full process of everything in the SSIS task.  The issue that I am having is that after that task completes, I need to go into Visual Studio to deploy the cube becuase we are unable to browse or use the cube.  This issue seemed to start once we changed the SSIS Analysis Service Processing Task to do a full process on the dimensions, rather than an incremental.

I would expect that once development is done, and it is processed and deployed, that is it.  My thinking is that the SSIS task should just update the already deployed cube,

View 2 Replies View Related

Analysis Services Processing Task

Apr 9, 2008

Hi. I have an Analaysis Services (2005) cube with four dimensions and one fact table (with three partitions - 2006,2007,2008) for which I need to create an SSIS package to process. I only want to process one of the three partitions (2008) - the previous two years should remain unchanged.


This is what I have currently in the Analysis Services Processing Task under Processing configuration:
- An object for each of the dimensions with "Process Full."
- An object for the 2008 partition with "Process Full."

(Note - Under Process Options, I see only Process Default, Process Full, Unprocess, and Process Data for dimensions and partitions).


Batch settings are:
- Processing order: Sequential
- Transaction mode: All in one transaction
- Dimension errors: Ignore errors
- Process affected objects: Do not process


When I execute the package, the cube loses the 2006 and 2007 data.

I am assuming that I have an issue with the Process Option or the Batch Settings, and I would appreciate any guidance!


Thanks,
Marianne

View 6 Replies View Related

Analysis Services Processing Task

Mar 11, 2008



Hi all, here is my problem:

The last node of my workflow in SSIS is an analysis Services Processing Task, which is supposed to fully reprocess a cube, defined in a different project.

In the configuration, I found the correct cube and setups for it, I thought I wasn't gonna have any problems with it, but it started to complain about user and password information. I thought since the databases configured itself when I added them, the same thing would happen with this Task.

I do have my own user and pass which has permissions to reprocess the cube, although I thought windows authentication would be better then setting up a user and password for the application/task.

I looked in the entire configuration pane and found no information regarding username and password.
Where should I set it up, my SSIS solution or the Cube's solution?

This might be a newbie question, I'm not quite sure...

EDIT: Here is the error message:
[Analysis Services Execute DDL Task] Error: The following system error occurred: Logon failure: unknown user name or bad password. .

View 5 Replies View Related

Analysis Services Processing Task

Feb 21, 2008



Hello

I am trying to run Olap Cube 2000 inside SSIS project.
I am using "Analysis Services Processing Task" Object.
The Visual Studio Project is sitting on the machine where the
analysis 2000 is running but yet i get an error while establish
a connection to the Analysis server.

On that machine also install MICROSOFT SQL SERVER 2005 .

the error is:
A Connection Cannot be made . Ensure that the server is running.

Does Anybody have an idea to why i get this Error.

Thanks,

View 1 Replies View Related

Where Is The 'Analysis Services Processing Task' Logging To

Apr 1, 2008

All,

I 'm using 'Analysis Services Processing Task' as part of a SSIS package to refresh the cube. in the property page,

the 'loggingMode' is set 'enabled', but there is no records in the sysdtslog90 table while all other tasks are logged in the table. How to logging into the sysdtslog90 table?



Thanks in advance

Jessie

View 3 Replies View Related

Error - Analysis Services Processing Task

Jul 11, 2007

We have an Integration services package that executes a few TSQL tasks, then processes an Analsys Services database. This has been in production for about three weeks now and twice the package has failed with this error from the event log:



Event Type: Error

Event Source: MSSQLServerOLAPService

Event Category: (289)

Event ID: 3

Date: 7/11/2007

Time: 1:48:59 AM

User: N/A

Computer:

Description:

OLE DB error: OLE DB or ODBC error: An error has occurred while establishing a connection to the server.

When connecting to SQL Server 2005, this failure may be caused by the fact that under the default settings SQL Server

does not allow remote connections.; 08001;

Communication link failure; 08S01;

TCP Provider: An existing connection was forcibly closed by the remote host.

; 08S01.

For more information, see Help and Support Center at http://go.microsoft.com/fwlink/events.asp.



I don't think that this error is accurate because the package and Analysis Services are on the same server.



Also, this does not happen in our development environment. Any help is appreciated.



Thanks,

Brian

View 1 Replies View Related

SSIS Analysis Services Processing Task - Big Problem!

Nov 1, 2007

I have an olap database "A" and SSIS package "P" which process all the dimensions and cubes in "A" olap database.

I created "A1" olap database copy of "A" and made copy of "P" SSIS package as "P1"
I opened "P1" SSIS package and updated olap connection properties "Initial Catalog = A1". A1 is my new olap database.

When I run package "P1" guess what? it processed "A" olap database's cubes and dimensions. Try it, not in production because I did it in production.

View 12 Replies View Related

Analysis Services Processing Task - Cube No Longer Updating

Oct 19, 2007



We have set up an IS package to process an AS 2005 database (comprising cube & dimensions, etc) daily, via a SQL Server Agent job on both development and production systems. This has been working fine for months.

A new dimension was added to the cube on the development system - automatic processing via the IS package continued without issue. However, when the new dimension was added to the production system the IS package no longer processes the cube correctly. Although all appears ok (and all is present and correct in the logs), no data updates to the cube are made. Only when the cube is manually processed does the cube get updated.

Anyone got any ideas about how to get around this issue? We have created a new IS package, with a single Analysis Services Processing Task, and tried this but get the same outcome.

View 4 Replies View Related

SSIS-Analysis Services Processing Task- On Client Fails

May 14, 2008

I am trying to execute an SSIS package from a client that contains an Analysis Services Processing Task in the package. The client that does not have SSIS and SSAS installed. We are getting an error

The task "Analysis Services Processing Task" cannot run on this edition of Integration Services. It requires a higher level edition. The same package runs from a server that has both SSIS and SSAS installed. Let me know if someone has come across the same problem.

Thanks

View 1 Replies View Related

Analysis Services Processing Task: Logging And Error Handling

Mar 5, 2007

I have an Analysis Services Processing Task in my SSIS package. I run the SSIS package using SQL Server job, the running of the package is a job step.

When I process manually the analysis services objects (in practise cubes) using dtexec utility I get a lot of log. In case the processing fails I get error messages that quite well describe the error. But when I run the job the only information I get in the job log is that the job step failed. I know the failure happens in the Analysis Services Processing Task.

Is there any way in SSIS to get a) the log of the Analysis Services processing or b) the error messages of the Analysis Services processing? Or should the processing be done some other way than I've been doing?

View 4 Replies View Related

Problems With Connections And Analysis Service Processing Task In SSIS

May 30, 2008

Hi everybody,

I'm fairly new to the SSAS/SSIS world (though not new on databases, etc.) and I'm having some problems with the SSIS packages in our Cube environment.

Currently in our SSAS/SSIS project, we have two major connection managers, one to the database we use for loading the Cube, and the other connector for the cube itself. To load the data from the database to the cube, we wrote some SSIS packages and used the Analysis Service Processing tasks to process all the dimensions and measures. This works pretty good, so no problems here.

The real problem starts, when I try to change the connection parameters, e.g. because the server changed, or the database has been renamed.
As soon as the connection managers points to another (existing) cube, regardless if the structure is exactly the same as the one of the old cube, the tasks lose all the assigned objects from their lists. It is really annoying to add all these exactly same objects to the task again. I tried experimenting with the DelayValidation attribute so the Development Studio doesn't destroy my work every time, but when I deploy the package the Cube breaks. Obviously some kind of deeper connection is destroyed when I change the connection string.

Is there a way to prevent the package from breaking/losing objects, without me having to sacrifice 15 minutes every time I change the connection parameters?

Regards,
Tris

View 4 Replies View Related

Changing 'initial Catalog' On Connection Causes Analysis Services Processing Task To Fail

May 2, 2008

We find that if we deploy the OLAP database with a different name on the test server, then regardless of how we change the connection string provided to the SSIS package that processes the cube, then the package fails to connect to the database. To clarify:

In development the OLAP database is called MyOlapDB and the source database is called MySqlDB. Both are on the same machine. When the the application is built and released for test, the test team install the databases on a replica of the production environment (i.e. web app on one machine, OLAP DB on another and SQL database on yet another). They also, quite rightly, implement the new test databases so they incorporate the build version number. So, MyOlapDB123 and MySqlDB123 are both from build 123.

This is when the problems start. Regardless of how the connection string is specified in the job that processes the cube, the SSIS integration package fails with the error:

[Analysis Services Execute DDL Task] Error: Errors in the metadata manager. Either the database with the ID of 'MyOlapDB' does not exist in the server with the ID of 'OurTestServer', or the user does not have permissions to access the object.


We have tried config files and job properties, but neither work. Also, simply attempting to run the package using the DTEXECUI does not work either.

Looking inside the XML of the package, we clearly see the ConnectionManager object which has the original connection string, which is

Data Source=localhost;Initial Catalog=MyOlapDB;Provider=MSOLAP.3;Integrated Security=SSPI;Impersonation Level=Impersonate;


However, editing the initial catalog here still does not solve the problem. Searching the XML for the string MyOlapDB reveals the OLAP database name in two other places - both within the object data of the two Analysis Services Execute DDL tasks.

Anyone know how to solve this problem without having to hack the XML of the package?

View 4 Replies View Related

Analysis :: SSAS Partition Processing

Oct 21, 2015

I am facing issue with partition processing. I am having a SSAS cube which is having 5 partitions. These partitions are processed through a sql server job using SSIS packages. In packages I used SSAS process task to do this.Now problem is, job is running successfully and showing that the step which is having partition process also fine.But data is not updating in the partition. While checking the partition properties, it is not updated with recent date and time.

When I try to manually process the partition, it is getting succeeded and recent data is getting reflected with recent date and time.Package configuration is done in job itself.

View 4 Replies View Related

Analysis :: Dimension Processing Error

Mar 14, 2013

I'm getting this error during processing one dimension.OLE DB error: OLE DB or ODBC error: SQL Server blocked access to STATEMENT 'OpenRowset/ OpenDatasource' of component 'Ad Hoc Distributed Queries' because this component is turned off as part of the security configuration for this server. A system administrator can enable the use of 'Ad Hoc Distributed Queries' by using sp_configure. For more information about enabling 'Ad Hoc Distributed Queries', see "Surface Area Configuration" in SQL Server Books Online.; 42000.my dimension contains member from two datasource table.

and in another dimension i get error;Errors in the high-level relational engine. The 'dbo_vicidial_Users' table that is required for a join cannot be reached based on the relationships in the data source view.Errors in the OLAP storage engine: An error occurred while the dimension, with the ID of 'Staging User', Name of 'DimUsers' was being processed.

View 4 Replies View Related

Analysis :: AttributeKey Not Found When Processing The Cube

Apr 21, 2015

I have designed a cube. It has two fact tables and some dimensions. Fact table to fact table is many to many relationship.

For example

FactMain
DataKey(PK), StartDateKey, PostCodeKey, TotalCost
FactBridge
DataKey(FK), ProductKey(FK), Position  - PrimaryKey on DataKey + ProductKey + Position
DimProduct
ProductKey(PK), ProductCode

Cube is built successfully, processed successfully.When I try to process the cube from agent job, I am getting error "Attribute key not found: tablename, value..." I have added a job step to run AnalysisServices Command. I have taken the command from cube process script(taken from manually process the cube and take script generated). I used ProcessAffectedObjects = "true" in the script. When I checked the tables, the key does exist. Why am I getting this error?

View 5 Replies View Related

Analysis :: Duplicate Attribute Key Has Been Found When Processing

Jul 8, 2014

1) Errors in the OLAP storage engine: A duplicate attribute key has been found when processing:

Table: 'dbo_Dim_x0020_Document_x0020_Type',Column: 'Item_x0020_No_', Value: '1100'. The attribute is 'Item No'.

How can I resolve this on package level.

2) I am also not able to see all the fields of a fact table when creating cube, where I can se all fields in dataview.

View 6 Replies View Related

Analysis :: Automate Execution / Processing Options

Oct 1, 2015

We plan to process our SSAS Cube nightly after our data warehouse is loaded (SSIS package) using an SQL Agent Job.

1. What is the best option to automate the processing of our cube? 
2. Can this be added to our SQL Agent Job?
3. As we will only be adding new dimensions and fact records, will be use Process Add?
4. Does the initial load require Process Full?
5. How can we configure a processing option before the automated execution?

View 2 Replies View Related

Analysis :: Processing Cube In Production But Not In Dev And QA Servers

Jul 30, 2015

I am getting following errors when i processed Cube in Production.An error occurred while the dimension, with the ID of 'DIM_PARTICIPANT', Name of 'DIM_PARTICIPANT' was being processed.
End Error

An error occurred while the 'PARTICIPANT NAME' attribute of the 'DIM_PARTICIPANT' dimension from the 'XL_GCS_SelfServices' database was being processed.
End Error

An error occurred while the dimension, with the ID of 'v d Transaction', Name of 'DIM_TRANSACTION' was being processed.
End Error

An error occurred while the 'INVOICE NUMBER' attribute of the 'DIM_TRANSACTION' dimension from the 'XL_GCS_SelfServices' database was being processed.
End Error.

But i implmented same in Dev and QA server, there is no issues found.

View 4 Replies View Related

Error In Processing Cube (Microsoft Analysis Services)

Jan 9, 2004

Hi,
I am processing one cube using Full Process option and it's giving
following error.

Analysis Server Error: Internal error [Object does not exist] '11948' ;
Time:1/8/2004 6:11:11 PM

Error(-2147221421): Internal error (Internal error [Object does not
exist] '11948' ); Time:1/8/2004 6:11:11 PM

Can anyone help me on this.

View 1 Replies View Related

Analysis :: SSAS 2012 - Notification While Cube Is Processing

Apr 16, 2015

I am working on SQL 2012.We have a SSAS Cube build. On top of it client use Excel to connect to SSAS Cube and See the reports.My Cube process every hour and take almost a 1-2 min to Process.When ever End user/Client See or refresh the Excel report (Which use Cube as Source),WHEN CUBE IS PROCESSING ,they get an error that Source is not available

We Have tried best but cannot bring down the Processing time of the cube to 3-5 Second , so that End user don't face report refresh issue at the moment of cube processing 

Requirement :In Case End User see the report while Cube is processing from back end , Instead of Error they should see some customize Msg which we can provide some thing Like "Data is Refreshing , please wait ".

View 5 Replies View Related

Analysis :: Processing SSAS Tabular Model Automatically

Oct 18, 2012

how can i automatically update my tabular model in the future when there's an update in my database.

View 4 Replies View Related

Analysis :: SSAS Tabular Model Processing Error

Nov 4, 2015

I get the following error while processing a SSAS tabular model (2014) on a new server.The SSAS service on this server is running under a login which has access to the SQL server data sources. I tried changing the provider to OLEDB from SQLCLNI11 in the connection string but that doesn't work too. The error message isn't useful to debug further.

The cube processing succeeds on a different server. I scripted out the cube DB and ran it on the new server and am trying to process full but it fails with the following error.

Error Message:
The operation failed because the source database does not exist, the source table does not exist, or because you do not have access to the data source.

More Details:
OLE DB or ODBC error: A network-related or instance-specific error has occurred while establishing a connection to SQL Server. Server is not found or not accessible. Check if instance name is correct and if SQL Server is configured to allow remote connections.

For more information see SQL Server Books Online.; 08001; SSL Provider: No credentials are available in the security package

; 08001; Client unable to establish connection; 08001; Encryption not supported on the client.; 08001.
A connection could not be made to the data source with the DataSourceID of 'd7a37dae-be87-44e0-a8b2-498069af82c9', Name of 'connection name'.
An error occurred while processing the partition 'XXXX_460f3467-1a99-4dc9-aaf2-bcf3d54a5c4c' in table 'XXXX_460f3467-1a99-4dc9-aaf2-bcf3d54a5c4c'.

The current operation was cancelled because another operation in the transaction failed. The operation failed because the source database does not exist, the source table does not exist, or because you do not have access to the data source.

View 3 Replies View Related

Analysis :: Processing / Deploying Adventure Works Tabular Cube

Sep 14, 2015

I am trying to create a Adventure Works Tabular cube in my developing machine and im having problems deploying the tabular cube (the DW database was firstly attached without problems)

So I'm having problems processing Adventure Works Tabular Cube, the problems are somewhere between SSAS project configuration and running privileges, i think. I'll describe exactly where i am:

- in SSMS 2012, connections, tables and roles and roles are empty, i only have the full name of the database when I try to process i get an error "the table with id of "product category 92583829......", Name of Product Category refereced by the model Cube, does not exist. An error ocurred when loading the model cube, '?c:program filesmicrosoft
sql serverMSAS11.MSSQLSERVEROLAPDataAdventureWorks Tabular model SQL 2012.0.dbModel.21.cub.xml'. (Microsoft AnalysisServices.)"

- in SQL Server configuration Manager, I have
Sql Server (MSSQLSERVER) running with account NT ServiceMSSQLSERVER
Sql Server Analysis Services (MSSQLSERVER) running with account .L.ABCDE , which is my account, i've changed because i've read somewhere that i could have problems with privileges.

(Note: it's strange because SSAS is tabular, i can see a blue rectangle (sort of) in SSMS but in configurator manager seems to be a cube.)

- in SSAS Project i unzip a clean copy of the project and the first error is that BIM file can't be found, it try's to find server TOSHIBA_S50_BTAB, but as i have only tabular SSAS running, i replace for TOSHIBA_S50... And some time ago it worked, now it doesn't even found TOSHIBA_S50_B?!!!!

View 3 Replies View Related

Analysis :: SSAS - Dimension Processing Failing Due To BitLocker Encryption

May 8, 2015

I'm facing an issue while processing OLAP. I have enabled BitLocker for dirve encryption and OLAP services uses this drive for db storage. OLAP is executing though SSIS package and I'm getting below error in Package. When debugging the script, it says Drive is encrypted using BitLocker. 

My client requires TDE for all databases, for OLAP we decided to use BitLocker: [URL] ....

SQL server is installed on C Drive & D Drive is the storage location for OLAP DB. When locking D Drive, OLAP processing failed. When I tried to restart SQL Server Analysis service in Services.msc it is not starting. Service restarted only when D Drive is unlocked. Is there any way we can process OLAP even when the drive is locked?

Error message is given below:

"The following system error occurred:  This drive is locked by BitLocker Drive Encryption. You must unlock this drive from Control Panel. "

View 3 Replies View Related

Analysis :: Processing Database Failed - Cannot Connect To Data Source

Apr 23, 2015

I have deployed my SSAS project to an Analysis services database on a SQL 2014 server instance using the "Development" build configration in Visual Studio. In order to avoid users accessing the development instance of the cube, I've created an additional "Release" build configuration and deployed my project to a different database on the same server. I've also created a production copy of the data source and changed the data source configuration of the production cube to point to this one.

I've provided the same domain service account on the "Impersonation information" tab of the data source for both the development as well as the production instance of the cube. This account has also been granted identical permissions for both data sources. While everything works fine with the development database, processing the production database fails with an error message saying that it cannot connect to its respective data source.

I'm using VS Ultimate 2013.4, SSDT 12.0.50318.0.

View 3 Replies View Related

Analysis :: Errors In OLAP Storage Engine - The Attribute Key Cannot Be Found When Processing

Jun 12, 2015

I 've read that there is a workaround for this issue by customizing errors at processing time but I am not glad to have to ignore errors, also the cube process is scheduled so ignore errors is not a choice at least a good one.

This is part of my cube where the error is thrown.

DimTime
PK (int)MyMonth (int, Example = 201501, 201502, 201503, etc.) Another Columns
FactBudget
PK (int)Month (int, Example = 201501, 201502, 201503, etc.)

I set the relation between DimTime and FactBudget doing DimTime MyMonth as Primary Key and FactBudget Month as Foreign Key.
The cube built without problem, when processing the errror: The attribute key cannot be found when processingwas thrown.

It was thrown due to FactBudget has some Month values (201510, 201511, 201512 in example) which DimTime don't, so the integrity is broken.

My actual question: is there a way or pattern to redesign this DWH to correctly deploy and process?

View 4 Replies View Related

Analysis :: Looping AMO Objects Versus Using SSAS Built-in Parallel Processing?

Jul 22, 2015

I am creating an SSIS Script Task that will be used to process SSAS dimensions and partitions and ideally log the details of each in a table. Any info on the benefits or drawbacks of using the built-in SSAS parallel processing as opposed to doing it manually in a multi-threaded "Parallel.Foreach" loop using the .NET AMO library.

In my testing, when I use a Parallel.foreach loop, I am able to obtain and log information about the object such as end time and time to process immediately after each object is processed.  This allows me to keep a history of processing time for each object:

public void processDimensions(Server Server, Database Database, ProcessType processType)
{
Parallel.ForEach(Database.Dimensions.OfType<Microsoft.AnalysisServices.Dimension>(), d =>
{
DateTime beginTime = DateTime.Now;
try
{
d.Process(processType);

[code]....

If circumventing the built-in SSAS parallel processing is not best practice I'd like to know in advance before I go too far down that path.

View 2 Replies View Related

Analysis :: Dynamic Values In XMLA To Incrementally Update Cube Processing

Nov 10, 2015

I am trying to incrementally update a Cube to get near real time data for the end users. Currently we have a Sql server agent Job that does a FullProcess on the Cube. The Cube consists of a single Measure group which is simply one named query containing inner joins of all the dimensions and fact tables in the underlying relational database. The end users have a lot to upload during the day and they would like us to refresh the cube (near real time) to ensure the adjustments are loaded so that they could reconcile their daily PnLs. We have a MeasureId added which is an auto increment column in the Measures table.

I am trying to schedule the below XMLA query in Sql server agent Job and schedule it to run every 15mins or even less (if possible). However it seems to be not working and keep throwing all sorts of errors.

DECLARE @LastMeasureId AS INT, @myXMLA nvarchar(max)
SELECT @LastMeasureId = "[Measures].[Maximum Measures Id]" FROM
OpenRowset(
'MSOLAP',
'DATA SOURCE=L68F728326574; Initial Catalog=GMDR;',
'SELECT NON EMPTY {[Measures].[Maximum Measures Id]} ON COLUMNS FROM [GMDR]');

[Code] ....

View 2 Replies View Related

AS Processing Task

Nov 27, 2005

When using the AS processing task with a connection to "an Analysis Services project in this solution", only some processing options are available for processing dimensions. For instance, it is not possible to select "Process Update". Once I change the connection manager to point to the deployed cube database, I can choose from all the options. Is this by design?

View 1 Replies View Related

Analysis :: SSAS (BISM) 2014 Full Processing Number Of Threads Are Limited

Jul 6, 2015

I have been tasked with processing a large tabular cube using SQL AS 2014 (with latest CUs).The three Fact tables having 1.2 billion rows (in each table)  have been divided into 30 vertical partitions to aid in parallel processing.  So around 40 million rows per partition.

Using SQL Profiler to monitor the Row counts (IntegerData) of records processed seems to max out around 2 million rows per minute, then tapers down to about 200k /minute.

The processing is taking over 14 hours and I need to get it lower if possible. The server has 48 cores (2.66MHz) and over 1TB RAM installed. But I really don't ever see CPU exceed 20% having a maximum of 206 threads running on the instance msmdvr.exe

Available RAM is always at least 30% (or 300GB).

I have increased the Vertipaq MIN/MAX 60%/80%

I have increased the OLAP / Processing / Max Thread Pool Min 500 and Max to 1000.

The connection properties have been increased to allow 100 connections, the majority of the processing consumes about 92 connections for the 90 large partition views for the facts.

What can be done to increased the server resource utilization and decrease processing times?

I have increased both

View 5 Replies View Related

OLAP Services Processing Task

May 24, 2002

Hi Guys.

For SQL 2000 is there any addin available for DTS task?

If not how can i automate it?
Advance thanks

-MAK

View 1 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved