How can I delete the file(sql text document)? When I tired to delete the file, the message is displayed 'There has been a sharing violation. The source or destination file may be in use.'
Does anyone know if it is possible to specify the size of a document map in a report? Ot it's scroll bar?
I have a report that contains about 100+ groups and therefore 100+ items in the document map. On initial display of the report, about 25 groups are visible in the document map, and you can see that there is more data below (half a row is shown) but the scroll bar does not allow scrolling all the way to the bottom of the list.
Out of Memory when working with big XML Files:when validating XSD it process small files but when size close to 1gb it throws .I have 16 xml files out of which 8-10 Â files size will be around 1gb processing one by one in FOR EACH LOOP container in SSIS[XML Task] Error: An error occurred with the following error message: "Exception of type 'System.OutOfMemoryException' was thrown.".Task XSD Validation failed.
system configuration : processor:Intel(R) Xeon (R) CPU E5-2670 v2 @2.50GhZ 2.50 Hz Installed memory (RAM): 61 GB System type: 64 -Bit operating System Visual studio:2012,32 bit virtual memory : 12499 MB
whether the size of the XML Document a limitation for the validation task ? or any system configuration i need to improve?
I'm designing a Job Recruitment Website, in which the admin person searches for the right candidate for the job using certain keywords . Each jobseeker will be uploading his CV (ms word doc) during registration .How can i search for keywords in the word documents. I just want the candidate reference once i found keyword match in the word docs.I heard about the indexing and blobs in sql server? But dont know much about it Are these the only solutions ?Is there any better approach for this ?Any help will be greatly appreciated
hi, guys Does anyone know how to change the Document Map root text? For example, i have report, the file name is sc.rdl, and then the root is sc. Apparently this is not good. I am thinking is there a way to change it.
Hi all,I've the following scenario: One Full Text Search enabled SQL Server Table with one image type field that holds document uploaded from users.The idea is to store binary document(.doc. xls, .pdf, .ppt, html, .xml and so on) and using SQL server full text search i can retrieve record that contains certaind word or words.I have no problems with them. So imagine i do a simple select * from Documents where a=b and i get one collumn with binary document field.With this scenario, i want to "extract" that text from that document + 20 words left and 20 right to show user some info and help him to search desired document (not only for its type or title). Like search engines do.But believe me, i cant find out any component, class or something to do such a thing.I think, the hard work is done with full text search engine...Sql server has that data, but it can decodeI'm so desperate. I would accept answers such "sorry, it can't be done", from a experienced user, but i need to know Thanks in advance
I've created a dataset with 27 measures and 20 query parameters. When attempting to load the report containing this dataset I'm shown the message;
'Document contains one or more extremely long lines of text. These lines will cause the editor to respond slowly when you open the file. Do you still want to open the file.'
If I do open the file it does indeed respond very slowly or even hangs.
I can manually format the XML code but amending the code in any way (i.e. using the layout designer to move a chart) removes my formatting and re-introduces the problem.
Are these an unreasonable amount of measures / parameters?
Environment; VS2005 v8.0.507 MSSQL 2005 9.00.1399.06 Build 3790 SP2 Windows Server 2003 SP2
I am working the Books Online documentation for the full-text search feature of SQL Server 2005 Express Advanced and having a problem following the instructions.
I made sure to choose the "Full Text Search" option during installation of VB 2005 Express Advanced.
I downloaded, installed, and attached the AdventureWorks database successfully.
I checked to ensure that the database was enabled for full-text search, but could not follow the instructions for indexing a table within the database. Here are the instructions from Books Online: To enable a table for full-text indexing
Expand the server group, expand Databases, expand User Databases, and expand the database that contains the table you want to enable for full-text indexing.
Right-click the table that you want to enable for full-text indexing.
Select Full-Text index, and then click Enable Full-Text indexing.
Another document notes:
To create a full-text index on a table, the table must have a single, unique not null column. For example, consider a full-text index for the Document table in Adventure Works in which the DocumentID column is the primary key column.
When I right-click the Document table (Production.Document) in the AdventureWorks database, there is no option to "Select Full-Text Index" or "Enable Full Text indexing".
Am I missing something here?
How do I get the the table indexed for full text search?
I have a Full Text index on a table with an image field that is successfully indexing .doc, .pdf and .rtf files.
Keyword searching this is no problem.
What i want to be able to do is perform a similarity search. by this i mean pass in a Key_ID (documentID) and have the database return a list of Key_IDs (documents) which are similar.
By similar i mean contain mostly the same keywords in roughly the same quantities
Is it safe to change 'max text repl size (B)' on a live system while it is running with users and services running?? I'm using SQL Server 7.0 running on Windows 2000.
Does anyone have experience with Full Text catalogs on large tables? We have a full text catalog on a table with about 30 million rows. Acording to BOL, once you get over 1 million you'll need to make some adjustements. Our system works, but we randomly get the following errors with ad hoc queries. The server has 8GB of RAM and 4 3GHz processors. Does anyone have experience working with a table this big? Any suggetions as to what could cause these errors? The only thing I could find was BUG#: 469483 on MS's site, but we're not using any OR clauses
Thanks
Here's the errors....
(query) SELECT * FROM acc_results ar WHERE CONTAINS(finding_text, 'cell')
#1 Server: Msg 7619, Level 16, State 1, Line 1 Execution of a full-text operation failed. Not enough storage is available to process this command.
#2 Server: Msg 7342, Level 16, State 1, Line 1 Unexpected NULL value returned for column '[FULLTEXT:acc_results].KEY' from the OLE DB provider 'Full-text Search Engine'. This column cannot be NULL. OLE DB error trace [Non-interface error: Unexpected NULL value returned for the column: ProviderName='Full-text Search Engine', TableName='[FULLTEXT:acc_results]', ColumnName='KEY'].
#3 Server: Msg 7399, Level 16, State 1, Line 1 OLE DB provider 'Full-text Search Engine' reported an error. [OLE/DB provider returned message: Not enough storage is available to process this command.] OLE DB error trace [OLE/DB Provider 'Full-text Search Engine' IRowset::GetNextRows returned 0x80004005: ].
The Upgrade Advisor noted that SQL Server 2005 changed the way full text catalogs are utilized. It says that you must ensure tha tthe file group associated with the base table has enough space to accomodate the additional space requirements for full-text indexes. It says to use the formula (2*FTK + 34 bytes) * RC where FTK=full text key size and RC=rowcount of the table. I know which table but if I do sp_help on the table, it doesn't tell me anything about the size of the full text key. I've tried searching TechNet and I'm either not finding the answer or I don't know enough to know if it's there. Can anybody help me?
I wrote this sql function which takes a comma seperated string of numbers, splits the numbers seperately and stores it in a table. I have specified the input parameter type as text instead of varchar, the size of the string can get more than 8000.
But the function is not working properly if the input size is more than 8000. For example if the input string is of length 8005 and this is the input string from 7995 to 8005 - '123,124,125'. It works fine till 123 and after that it throws an error, Syntax error converting the varchar value '124,125' to a column of data type int. Can anyone tell me what is wrong with this. I am using string functions like charindex, substring. I can post the full function if you want.
I have a matrix with a title column. My matrix has a column that is expanded in 2 with data changes as followings:
MY TITLE1 MY TITLE2 YEAR 2007 YEAR 2008 } matrix matrix Column column } matrix
How can I change de title (MY TITLE2) differently according to the changing value of the dynamic data column ? My title2 has to be located above the YEAR 2008 and has to be different from MY TITLE 1.
Thank you for your help. Best Regards, Hirondelle.
In a title of a column of a table, I put a too long text, with sorting, but when I spread (display) the report has a problem on text of the title of the column there.
The probleme settles (arises) when I activate the sorting to post (show) the image of sorting.
I have a problem I can't seem to find the solution with this aweful limitations on VARCHAR fields of 255.
Within a stored procedure called Store_Check, I need to dynamically build a string (@string) using VARCHAR(255) since the text datatype can't be used in a stored procedure.
So,this string is built according to whether the Store ID is NOT NULL. So if the StoreID is not null, I start building this string 'Exec Update_Store_Address @StoreID1, @address2'. There are 20 StoreID's passed into Store_Check. IF all 20 StoreID's are not NULL, the executed String greatly exceeds 255 because the string winds up looking like this
I am not executing this string within the StoredCheck procedure. It needs to be passed as ONE string to a VB program and it gets executed by the VB program. Even if I create 4 local variables and concatenate them, it stops at the 255th character.
Also, a local varialbe of type TEXT cannot be declared within stored procedure.
Hello,A while back I dropped a text column from a SQL Server 7 databaseroughly 3GB in size. I expected the size of the database to decreaseby around 1GB, but no change occurred. After searching usenet, Idiscovered that SQL Server 7 has no way of reclaiming that space, butthat there is some command that can be run in SQL Server 2000 thatwill reclaim it.I have since migrated this database to SQL Server 2000, and am nowtrying to figure out what that command is, but cannot locate anyusenet posts about it... also tried searching books online, but can'tfind anything that way either.Does anyone know what I should run?Thanks,Tom
I received a pipe-delimited file that I need to import. (It has the equivalent of 650+ fields on a single row). While I had no issue importing it (SSIS 2008) I noticed that the input connector, Advanced option, shows an "OutputColumnWidth" of only 50 for all fields.
I say only 50 because some of the pipe-delimited fields can supposedly have a max of 250 characters so I'm concerned about potential data truncation. Unless someone has another thought I plan to manually set those OutputColumnWidth fields to 250.
I am trying to resize a database initial log file from 500M to 2M. I€™m using€?
ALTER DATABASE <DBNAME> MODIFY FILE ( NAME = <DBLOGFILENAME, SIZE = 2 ) "
And I'm getting "MODIFY FILE failed. Specified size is less than current size." I tried going into the database properties and setting the log file to 2M, but it doesn€™t keep the changes.
I installed sql 2005 a while back. Then I recently found out my file system was fat32 (I don't understand why the hardware people did this...) and I had to convert to NTFS. Naturally the sql service no longer worked so I uninstalled inorder to reinstall now I can't reinstall it I keep getting this message
native_error=5039, msg=[Microsoft][SQL Native Client][SQL Server]MODIFY FILE failed. Specified size is less than current size.
Hi folks,Can anyone enlighten me here? I'm trying to use a SPROC which, when supplied with an int, looks up the table and returns certain columns from it. I'm using a SqlCommand, here's my codebehind: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ SqlCommand dataSource = new SqlCommand("retrieveData", new SqlConnection(dbConnString)); dataSource .CommandType = CommandType.StoredProcedure; dataSource .Parameters.AddWithValue("id", poid); dataSource .Parameters.AddWithValue("title", title).Direction = ParameterDirection.Output; dataSource .Parameters.AddWithValue("creator", creator).Direction = ParameterDirection.Output; dataSource .Parameters.AddWithValue("assignee", assignee).Direction = ParameterDirection.Output; etc, etc... And the SPROC:------------------------------------------------------------------------------------------------------------------set ANSI_NULLS ONset QUOTED_IDENTIFIER ONGOALTER PROCEDURE [dbo].[retrieveData] @id int, @title varchar(50) OUTPUT, @creator varchar(50) OUTPUT, @assignee varchar(50) OUTPUT, @contact varchar(50) OUTPUT, @deliveryCost numeric(18,2) OUTPUT, @totalCost numeric(18,2) OUTPUT, @status tinyint OUTPUT, @project smallint OUTPUT, @supplier smallint OUTPUT, @creationDateTime datetime OUTPUT, @amendedDateTime datetime OUTPUT, @locked bit OUTPUT AS /**SET NOCOUNT ON; **/ SELECT [title] AS [@title], [datetime] AS [@creationDateTime], [creator] AS [@creator], [assignee] as [@assignee], [supplier] as [@supplier], [contact] AS [@contact], [delivery_cost] AS [@deliveryCost], [total_cost] AS [@totalCost], [amended_timestamp] AS [@amendedDateTime], [locked] AS [@locked] FROM purchase_orders WHERE [id] = @id; ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ The id being passed in is definately not null, and is set to a value of an item I know exists. The resulting error is:
Exception Details: System.InvalidOperationException: String[1]: the Size property has an invalid size of 0.Line 63: retrievePODetails.Connection.Open();Line 64: retrievePODetails.ExecuteNonQuery();[InvalidOperationException: String[1]: the Size property has an invalid size of 0.] System.Data.SqlClient.SqlParameter.Validate(Int32 index) +717091... ... Can anyone see anything I'm missing? Thanks,Ally
Using C#, SQL Server 2005, ASP.NET 2, in a web app, I've tried removing the size from parameters of type NCHAR, NVARCHAR, and VARCHAR. I'd rather just send a string and let the size of the parameter in the SP truncate any extra chars if need be. I began getting the error below, and eventually realized it happened only with output parameters, as in the code snippet below.String[3]: the Size property has an invalid size of 0. par = new SqlParameter("@BusinessEntity", SqlDbType.NVarChar); par.Direction = ParameterDirection.Output; cmd.Parameters.Add(par); cmd.ExecuteNonQuery();What's the logic behind this? Is there any way around it other than either finding out what the size should be, or assigning a size larger than would ever be needed? ThanksMike Thomas
I have one db test with one .mdf and .ldf file...mdf file size is 100mb and for some reson i removed all the tablesfrom that .mdf file and transfer it into new secondary file so all thetables moved into secondary file now i want to reduce the first .mdffile from 100 mb to 50mb is that possible,it's showing 90mb is free.Please reply
I am getting error to run stored procedure using executenonquery method. The Stored Procedure is having OUTPUT parameter. ExecuteNonQuery statement is called using SqlHelper. Error : String[18]: the Size property has an invalid size of 0
Just wanted to know what is a general rule of thumb when determining log file space against a database's data file.We allow our data file for our database to grow 10%, unlimited. We do not allow our log file to autogrow due to a specific and poorly written process (which we are in a three month process of remove) that can balloon the log file size.Should it be 10% of the Data file, i.e. if the Date file size is 800MB the log file should be 8MB?I realize there are a myraid of factors that go against file size but a general starting point would be nice.ThanksJeff--Message posted via http://www.sqlmonster.com
I have a matrix table which has two groups in the columns
I am trying to create a multileveled document map which i have succeed by
1. Edit Group 1 properties and setting the document map properties to the name of the field 2. Edit Group 2 properties and setting the document map properties to the name of this field and making the parent group the name of Group 1
The document map appears with the correct groups and name , however when I click on the values it doesnt take me to any where !!! It refreshes the screen but it doesnt go anywhere !
Can someone please help me troubleshoot this or know what i may be missing