SQL Server Destination Not Loading All Rows

Jan 23, 2008

I have a SSIS package that transfers data from three SQL Server 7 servers to a SQL Server 2005 database. This package has about 30 different tables it copies. The table structures in the source database and destination tables are identical. About 25 different tables load without any issues. I have about 5 tables that load some nights without a problem. On other nights, the data transfers seem to randomly (though usually the most recent records) ignore some of the data. I have logging turned on and receive no errors. It just appears to stop loading data.

I should also mention that I truncate each destination table before begining and each table is loaded from data from each of the 13 source database (I am combining data from 13 regional database for reporting purposes). This is done using a Foreach Loop Container that updates the Server/Region connection string for each region. I am using the OLE DB Source connection to the SQL Server Destination. I have tried as well with the OLE DB Destination with the same result (and no error). I do not do any manipulation to the data on the transfer, but added a "RowCount" transformation between the source and destination and it gives the correct number of rows, but not all the rows get loaded.

View 6 Replies


ADVERTISEMENT

Integration Services :: Loading Flat Files Without Duplicate Rows Into Destination Server

Sep 25, 2015

I have some duplicate records in my flat file. But i don't want to load those duplicate rows into my destination.

View 2 Replies View Related

Any Way To Check The Duplicated Rows In Destination Before Loading Data?

Jun 5, 2006

Hi. As the title, I am try to figure out how to write script to prevent duplicated rows before loading data from couple csv files to the OLE database table.
Another quick question, when I use Data Conversion to convert data from string to datetime or decimal type, it always return error like potential data loss.

View 4 Replies View Related

Errors Loading A Text File Into A Sql Server Destination

Apr 24, 2006

I am trying to load 14+ million rows from a text file into local Sql Server. I tried using Sql Server destination because it seemed to be faster, but after about 5 million rows it would always fail. See various errors below which I received while trying different variations of FirstRow/LastRow, Timeout, Table lock etc. After spending two days trying to get it to work, I switched to OLE DB Destination and it worked fine. I would like to get the Sql Server Destination working because it seems much faster, but the error messages aren't much help. Any ideas on how to fix?

Also, when I wanted to try just loading a small sample by specifying first row/last row, it would get to the upper limit and then picked up speed and looked like it kept on reading rows of the source file until it failed. I expected it to just reach the limit I set and then stop processing.

[SS_DST tlkpDNBGlobal [41234]] Error: An OLE DB error has occurred. Error code: 0x80040E14. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Cannot fetch a row from OLE DB provider "BULK" for linked server "(null)".". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "The OLE DB provider "BULK" for linked server "(null)" reported an error. The provider did not give any information about the error.". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Reading from DTS buffer timed out.".

--------------------------------
[SS_DST tlkpDNBGlobal [41234]] Error: The attempt to send a row to SQL Server failed with error code 0x80004005.
[DTS.Pipeline] Error: The ProcessInput method on component "SS_DST tlkpDNBGlobal" (41234) failed with error code 0xC02020C7. The identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running.
...
[FF_SRC DNBGlobal [6899]] Error: The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.

[DTS.Pipeline] Error: The PrimeOutput method on component "FF_SRC DNBGlobal" (6899) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.

[DTS.Pipeline] Error: Thread "WorkThread1" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.


-------
After first row/last row (from 1 to 1000000) limit is reached:
[SS_DST tlkpDNBGlobal [41234]] Error: An OLE DB error has occurred. Error code: 0x80040E14. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Cannot fetch a row from OLE DB provider "BULK" for linked server "(null)".". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "The OLE DB provider "BULK" for linked server "(null)" reported an error. The provider did not give any information about the error.". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80040E14 Description: "Reading from DTS buffer timed out.".

---------------
When trying to do a MaximumCommit = 1000000. Runs up to 1000000 OK then slows down and then error.
[SS_DST tlkpDNBGlobal [41234]] Error: Unable to prepare the SSIS bulk insert for data insertion.

[DTS.Pipeline] Error: The PrimeOutput method on component "FF_SRC DNBGlobal" (6899) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.

----
When attempting all in a single batch:
[OLE_DST tlkpDNBGlobal [57133]] Error: An OLE DB error has occurred. Error code: 0x80004005. An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80004005 Description: "The transaction log for database 'tempdb' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases". An OLE DB record is available. Source: "Microsoft SQL Native Client" Hresult: 0x80004005 Description: "Could not allocate space for object 'dbo.SORT temporary run storage: 156362715561984' in database 'tempdb' because the 'PRIMARY' filegroup is full. Create disk space by deleting unneeded files, dropping objects in the filegroup, adding additional files to the filegroup, or setting autogrowth on for existing files in the filegroup.".

View 11 Replies View Related

To Validate # Of Excel Rows(source) And Sql Server Table(destination) Rows Are Equal

Feb 20, 2008

Hi,

When expoting data from excel to sql server table, using SSIS package, after exporting is done, how would i check source rows are equal to destination rows. If not to throw an error message.

How can we handle transactions in SSIS
1. when some error/something happens during export and the # of rows are not exported fully to destination, how to rollback the transaction in SSIS.

Any sort of help would be highly appreciated.

Thanks,

View 2 Replies View Related

To Validate # Of Excel Rows(source) And Sql Server Table(destination) Rows Are Equal

Feb 20, 2008

Hi,

When expoting data from excel to sql server table, using SSIS package, after exporting is done, how would i check source rows are equal to destination rows. If not to throw an error message.

Any sort of help would be highly appreciated.

Thanks,

View 1 Replies View Related

Loading From Oledb To Excel Destination

Mar 13, 2008



Hi,

I have a simple loading to excel destination. It has 900,000 records. In 66,000+ records, i has an error



Error: 0xC0202009 at Data Flow Task, Excel Destination [3286]: SSIS Error Code DTS_E_OLEDBERROR. An OLE DB error has occurred. Error code: 0x80004005.

Error: 0xC0209029 at Data Flow Task, Excel Destination [3286]: SSIS Error Code DTS_E_INDUCEDTRANSFORMFAILUREONERROR. The "input "Excel Destination Input" (3297)" failed because error code 0xC020907B occurred, and the error row disposition on "input "Excel Destination Input" (3297)" specifies failure on error. An error occurred on the specified object of the specified component. There may be error messages posted before this with more information about the failure.

Error: 0xC0047022 at Data Flow Task, DTS.Pipeline: SSIS Error Code DTS_E_PROCESSINPUTFAILED. The ProcessInput method on component "Excel Destination" (3286) failed with error code 0xC0209029. The identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running. There may be error messages posted before this with more information about the failure.

Error: 0xC0047021 at Data Flow Task, DTS.Pipeline: SSIS Error Code DTS_E_THREADFAILED. Thread "WorkThread0" has exited with error code 0xC0209029. There may be error messages posted before this with more information on why the thread has exited.

Error: 0xC02020C4 at Data Flow Task, OLE DB Source [1]: The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.

Error: 0xC0047038 at Data Flow Task, DTS.Pipeline: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED. The PrimeOutput method on component "OLE DB Source" (1) returned error code 0xC02020C4. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing. There may be error messages posted before this with more information about the failure.

Error: 0xC0047021 at Data Flow Task, DTS.Pipeline: SSIS Error Code DTS_E_THREADFAILED. Thread "SourceThread0" has exited with error code 0xC0047038. There may be error messages posted before this with more information on why the thread has exited.



I assume that this is caused by the no. of records im loading since the .xls file can only contain 65,000 records.... i tried using the .xlsx file but i guess it only accepts .xls.

What's the alternative to load to excel 2005 with this numbers of records?

thanks.

cherriesh

View 1 Replies View Related

Dynamic Destination Flat File Loading

Jan 4, 2006

Here's what I want to do -

Dynamically load a flat file from a dynamic source table-

The source table metadata is known via the SYSOBJECTS and SYSCOLUMNS tables- I can pull the column names, type and lengths from these tables based on the table name. (my goal is pretty simple- pull data from a table in our database, and save it down to a flat file)

Would this be enough to dynamically create the destination flat file?  If so, how do I do it?

Thanks

-rob

View 3 Replies View Related

Integration Services :: Data Flow Task Failed After Loading 29000 Rows Out Of 234567 Rows

Oct 13, 2015

I am facing an issue that Data flow task failing after loading 29000 rows out of 2lakhs rows.

I am loading data from .csv file to OLE DB Destination.

This data flow task is placed inside For each loop container.

is this issue because of any performance issue in SSIS packages such as buffer size.

find the error below:

DFT Load Data from FlatFile:Error: The conditional operation failed.
DFT Load Data from FlatFile:Error: SSIS Error Code DTS_E_INDUCEDTRANSFORMFAILUREONERROR. 

The "DER Add Calc Columns" failed because error code 0xC0049063 occurred, and the error row disposition on "DER Add Calc Columns.Outputs[Derived Column Output].Columns[M_VALUE_NUM]" specifies failure on error. An error occurred on the specified object of the specified component.  There may be error messages posted before this with more information about the failure.

DFT Load Data from FlatFile:Error: SSIS Error Code DTS_E_PROCESSINPUTFAILED.  The ProcessInput method on component "DER Add Calc Columns" (48) failed with error code 0xC0209029 while processing input "Derived Column Input" (49). The identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running.  There may be error messages posted before this with more information about the failure.

[code]....

View 8 Replies View Related

Problem On Loading Data To Db2 Destination Using Oledb In Ms-ssis.

Jan 1, 2007

hello

i am performing the ETL on the as400 db2 database using ms- dts,ssis.

i have built the connection b/w as400 and source to extract data from as400 to staging means in dataflow . when i have built the oledb connction for loading data to destination as oledb destination.then it will connct successfully to the db2 as destination but when execute the task then it not load data , and give provider error.

what can be good solution for this.

can u solve it.

rep plz.



View 2 Replies View Related

SQL Server 2008 :: SSIS Package Loading More Than 25000 Rows Using Excel ACE 12 Driver

Jun 1, 2015

I have a SSIS package running well in production however sometimes the package will fail when the excel file contains more than 25000+ rows.

The SSIS package is run by SQL Server Agent and is set to run in 32bit mode.

I checked the data by loading in batches and all data loaded successfully. But the funny thing is when I run the same package on my local development PC using BIDS and the same data file. The package loads all 25000+ rows successfully.

Is there some setting that is preventing all rows loading in the server environment.

View 4 Replies View Related

Integration Services :: Loading Data To Destination With Foreign Key Relationship

Apr 22, 2015

I have to load data into destination table, it has foreign key relation to two different tables called person table and organization table . sample data to be loaded is like

person_id organization_id
1                   Null
2                    NULL
Null                1
null               null

person table and organization table doesn't have null values in them, when I try to load this data none of them are laoded, I know either person_id or organization id having null value is failing foreign key constraint. But I want to transfer all the rows except  the ones having both nulls. how this can be achieved ?

View 7 Replies View Related

Integration Services :: Date Timestamp Not Loading Correctly Into CSV File As Destination

Nov 5, 2015

I have a simple package to load data from sql server db into a flat file. I have a date field in the source data base (data type DATETIME) when i open the csv file some show the exact time stamp and some records show just the seconds like (00:00:0.7). I used CAST CONVERT bu still the same issue.

AppliedDate
00:00.6
00:00.6
10/2/2015 0:00
10/2/2015 0:00
00:00.3
00:00.3

View 9 Replies View Related

Which Task Is Best For Loading 1 Million Rows

Mar 24, 2008

Hi,

i have to load 1million rows from database( or flatfile) to the database(or flat file).
which task is used as the best solution for this?

Appreciate any assistance in this regard.

Thanks,
Das

View 5 Replies View Related

Rows Unexpectedly Skipped While Loading Flat File

Nov 17, 2007

I tried to load a fixed width flat file with around 300,000 rows. However, only the first 8xxxx rows were loaded to the destineation table and the rest row were loading blank records. There was no error message showing during package execution. I've tried to split the file in half and the result was the same. So it wasn't the data file problem.

Would there be any buffering issue I need to cater for inside the package? Thanks!

View 10 Replies View Related

OLE DB Destination - Redirecting Rows

Feb 14, 2008



I'm getting some strange error handling behaviour when I tried to redirect rows when an error is encountered at the OLE DB Destination. I enabled the fastload option and set the maximum insert commit size to 10,000. My package is trying to insert 1,000,000 rows. When the row containing bad data is encountered, the entire batch is redirected to a separate table with no table constraints. So I have approx 990,000 rows of good data inserted, and approx 10,000 rows in an error table. However, only 2 of the 10,000 rows are actually bad data.

I changed the filter in the data source to reload 5 rows from the 10,000 rows of bad data. The 5 rows include the 2 rows of bad data. The fastload option and max insert commit size at the OLE DB Destination are not changed. When the package completed, 3 rows were inserted to the destination and the 2 bad rows are redirected again. Even though this is the outcome I wanted (only the real bad data is redirected), shouldn't all 5 rows be redirected since they should all be treated as a single batch?

My second question is, is there a way to get the best of both worlds - be able to load lots of data as fast as possible and only redirect the rows that are actually bad?

Thanks.

View 1 Replies View Related

The Order Of Insertion Of Rows Into Destination Is Not Same As The Order Of Incoming Rows

Dec 21, 2006

Hi ,

i am dealing with around 14000 rows which need to be put into the sql destination.,But what i see is that the order of the rows in the desination is not the same as in the source,

However it is same for smaller number of rows.

Please help ...i want the order to be same.

View 4 Replies View Related

Loading Data From Multiple Rows Into Single Row In Excel Sheet

Jan 9, 2008


Hi,

I want to load data into Excel file with following format,





Country

State

Total

Location


ABC

A

20

X1


30

Y1


C

100




XYZ

X

40



Basically I want to insert records from multiple rows into a single row; how can I achieve this using SSIS.
I am using Excel as a data source.

Any help is appreciated.

Regards,
Omkar.

View 8 Replies View Related

Delete Excel Destination Rows

May 15, 2008

Hey all

I am exporting table rows (based on a query) into an excel file but I don't want to append to the file. I would like to delete the rows that were previously added and then add the new data. The file has column headings and I would like these to exist all the time.

I know how to export the data but don't know how to delete 'old' data rows from excel.

Any guidance will be highly appreciated.

Many thanks,

Rupa

View 5 Replies View Related

SSIS Oledb Destination Not Writing Any Rows

Nov 20, 2007



I have a Dataflow task with oledb source that is using SqlCommand to retrieve data and oledb destination to write the source output to a table. I have access to both the source and destination databases.

The problem is the destination component is not writing any rows to the destination table eventhough the Source component is returning rows (I can see them in the Preview and the source database table as well).
I'm using "Table/View Name from Variable" for destination.

The Package executes without any errors but there is no output.

Any ideas?

Thanks.

View 7 Replies View Related

Filtering Rows That Already Exist In The Destination Table

Feb 17, 2006

Hi All,

What is the most straighforward way of not importing rows that already exist in the destination table (as determined by a primary key value)? Thanks.

Regards,

Daniel

View 3 Replies View Related

Integration Services :: Add Only New Rows To A Destination Table

Jul 22, 2015

How do I add only new rows to a destination table (when copying a table from another database every night) ?Every night I am copying a number of tables from one database to another.I only want to insert news rows (that are not in the destination table, but are in the source table) to the destination table.I might normally drop the destination table and just copy over the whole table, but in this case rows can be deleted from the source table, but I want to keep these old rows in the destination table (to maintain history). So I only want to add in rows to the destination table that have been added to the source table since last time.I guess I could copy the whole of the source table to a temporary table in the warehouse, then use a T-SQL merge command to compare and just add new rows to the destination table- but suspect that this is not the best way.

View 8 Replies View Related

Custom Script Destination - Wrote 0 Rows Problems

Jun 12, 2007

I have created a custom script destination using an article found on TechNet. The script executes fine and completes, but with one small problem. The last package execution log entry states "wrote 0 rows". This is a problem because I use this number to verify that all of the database records have been properly transferred.



I have monitored that database when the script is executing and have verified that the data is being transferred property.



Is there some function that I should be calling with each row prcessed in the custom script destination to "tally" the row count? This way the script component will display the proper row count.



Any suggestions would be appreciated.

View 7 Replies View Related

How To Feed Specific Rows From Excel Spreadsheet To OLE DB Destination

Jan 2, 2008

Hi,

I am in a bit of a quandry. I have an Excel spreadsheet source, but I only want certain rows from the spreadsheet, which will be delineated by a header and footer row (Start and End). Here is my script thus far (see below).

My question is, what is the correct way to do this? I guess I am stuck on syntax. i.e.) Do I need to create an OUTPUT BUFFER?

Thanks

==============================

Sample data:
blah
Start
Frank 1234 1234 1234
Sue 2345 2345 2345
End
blah
blah

==============================


Public Overrides Sub Input0_ProcessInputRow(ByVal Row As Input0Buffer)

Static Section As String


'Set sections

If Row.Name = "Start" Then


Section = "Start"

End If

If Row.Name = "End" Then


Section = "End"

End If


If Section = "Start" Then

'GRAB THESE ROWS AND SEND THROUGH THE PIPELINE
'IGNORE ALL OTHER ROWS!
'What syntax goes here?
'Output buffer?



End If


End Sub

View 9 Replies View Related

Custom Destination Component Logging - Wrote 0 Rows

Mar 17, 2006

I wrote a custom destination component. Everything works fine, except there is a logging message that is displayed that I cannot get rid of or correct. Here is the end of the output of a package containing my component:

Information: 0x40043009 at Data Flow Task, DTS.Pipeline: Cleanup phase is beginning.
Information: 0x0 at Data Flow Task, MyDestination: Inserted 40315 rows into C: empfile.txt
Information: 0x4004300B at Data Flow Task, DTS.Pipeline: "component "MyDestination" (9)" wrote 0 rows.
SSIS package "Package.dtsx" finished: Success.

I inserted a custom information message that contains the correct number of rows written by the component. I would like to either get rid of the last message "... wrote 0 rows", or figure out what to set to put the correct number of rows into that message.

This message seems to happen in the Cleanup phase. It appears whether I override the Cleanup method of the Pipeline component and do nothing, or not. Any ideas?

public override void Cleanup()

{

ComponentMetaData.FireInformation(0, ComponentMetaData.Name,

"Inserted " + m_rowCount.ToString() + " rows into " + m_fileName,

"", 0, ref m_cancel);

base.Cleanup(); // or not

}

View 6 Replies View Related

Losing Rows From File To Destination Table - Need Troubleshooting Help

May 30, 2006

I am losing data from time to time and can not figure out where the rows are vanishing to. In essence, I have 5 files that I process. The processing occurs on a daily basis by one SQLAgent job that calls 5 individual packages, each doing a different piece of work.

I stage each file to a temp table, then perform some minor transformations, then union the rowsets from each file before landing to another temporary table. A subsequent package performs additional transformations that apply to the entire dataset, before inserting them to their final destination. Along the way, I reject some records based on the business rules applied. Each package in the entire Job is logging(OnError, TaskFailed, Pre and Post execute). Their are no errors being generated. No rows are being rejected to my reject tables either.

Without getting into the specific transforms, etc. being used in this complex process, has anyone seen similar unexplained behaviour? I've not been able to identify any pattern, except that it is usually only 1 or 2 of the record types (specific to a source file), that will ever not be loaded. No patterns around volumes for specific source files. There are some lookups and dedupes, however I've seen the records 'drop off' before reaching these transforms. I have noticed I had my final desination load not using fast load. However sometimes the records disappear before even getting to my final staging table which is inserting using fast load. I am going to turn on logging the pipelinerowssent event. Any other suggestions for troubleshooting/tracking down these disappearing records?

Thanks

View 3 Replies View Related

Problems With Deleting Rows In Destination Table When Exporting To Excel

Mar 29, 2006

Hi,

I export one table to Excel using Export Wizard. Second time I have used this creator I have checked option "Delete rows in destination table" because table already exists in Excel file.

During execution I get an error:

"Error 0xc002f210: Preparation SQL Task: Executing the query "DELETE FROM 'MyTable'" failed with following error: " and here is my translation from my language: "Deleting data from linked table is not supported by this version of ISAM"

I have also saved this package to a file and here is the full connection string to my Excel file:

Provider=Microsoft.Jet.OLEDB.4.0;Data Source=C:Documents and SettingsprzemekPulpitmyfile.xls;Extended Properties="EXCEL 8.0;HDR=YES";

Few days ago I have been using SQL2000 on the same machine and the same export was working properly.

Any suggestions?



Thanks.

Przemo

View 1 Replies View Related

Only Create Destination Flat File When There Are Rows Returned From Ole DB Source

Oct 25, 2007

Hi,

I have a data task with the following requirements:
1) Run query against database to retrieve rows
2) Add header and footer row to the result set. The footer row must contain a count of the records.
3) Write the rows to a fixed width file if there were any data rows

I have got to the point that I can create the file (using a set of tasks that includes derived columns, sorts, aggregation and merges). However the file is created regardless of whether there were data rows returned.
I can't check the row count before proceeding as this isn't set until the data task ends. And if I try to split them into separate data tasks (so that I can access this variable and perform conditional execution) it becomes harder to access the original rows.

Do you have any recommendations on the best way to achieve this?
It all seems to be very complex and I'm starting to feel that it would be easier to do this outside of SSIS...
Please help me to keep the faith!

For those interested this is a slightly simplified version of what I have so far (all within a single data task):

1.Run dummy sql to create header row 2.Run main SQL to retrieve rows
| |
| 3.Multicast
| | |
| | 4.Create footer row by doing sum() in aggregate task
| | |
| 5.Merge body and footer
| |
6. Merge header with body and footer
|

7. Create fixed width file

View 5 Replies View Related

After SSIS Package Runs All Rows, All Fields Are NULL In Destination Table ???

Jul 21, 2007

I am copying a simple table from a Sql Server 2005 database to an *.sdf mobile database.

I am brand new to SSIS and I am probably doing something wrong. But after executing the SSIS package all the rows and all the fields are NULL in the destination database. I put a datagrid viewer between the OLE DB Source and the Sql Server compact edition destination and I can see the real data which is obviously not ALL NULL.



Does anyone have a clue as to why it would be doing this?





Any help would be much appreciated.



Thanks...



View 1 Replies View Related

OLE DB DESTINATION And SQL Server Destination

Jul 4, 2007

Hey All:



I was totally confused.

When designing the SSIS dataflow part, firstly , i tried SQL Server Destination because my target server is a sql server.

then execute the task with failure.

Then i tried to use OLE DB DESTINATION instead of SQL Server Destination.

This Dataflow worked.



i can not figour out why.

By the way , i used the connection is OLE DB.And i choosed OLE DB source as the datasource cuz i can not find SQL server datasource.



Who can tell me some reasons for this?



View 9 Replies View Related

Loading To SQL Server

Jan 26, 2006

What is the fastest way I can load data into SQL SErver -- it seems to me it's with Bulk Copy and minimal logging. But that seems very slow when I look at Orace's Data Pump and DB2's LOADer. Does SSIS help here -- from what I can tell, you help me tons and tons on the ET part, but the L is what I'm wondering about...

View 5 Replies View Related

DTS LOADING FROM MainFrame To SQL Server

Sep 18, 2000

hi
I am loading from Mainframe to Sqlserver.
Right now i am truncating the tables in sqlserver connecting to mainframe and downloading the tables to Sql server. The problem is sometimes i fail to connect to Mainframe. So is there any way of checking whether i can connect to mainframe and if succesful truncate the tables.
Thanks

View 1 Replies View Related

Loading Huge XML To SQL Server 2005

Jan 13, 2008

Hi,
 I am using VS2003 and SQL Server 2005. I will get the huge size xml from the web service. I need to load this xml to sql server 2005 table as records.
I got below solution from microsoft site. Load XML to Dataset and from Dataset to SQL Server 2005.
Note   If you call ReadXml to load a very large file, you may encounter slow performance. To ensure best performance for ReadXml, on a large file, call the DataTable.BeginLoadData method for each table in the DataSet, then call ReadXml. Finally, call DataTable.EndLoadData for each table in the DataSet.
Any other better solution is there to load huge XML to SQL server 2005?

View 1 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved