Extremely thanks for the reply . I am using transactional replication for the database . I will try with snapshot replication as you suggested .
You mentioned that it will work with Transactional replication only if the application uses the option ' with log ' for those transactions .
Can you let me know where can i set this option for transaction replication ? I am sorry but i am not well versed with database replication procedures and management .
Currently im developing an online Q&A "board" in ASP.Net 2.0. Where users can post a question of a specific category and answers of a specific questions.
For this Q&A board i need a database to store data but i have no idea how the database design should look like for this. So i hope you guys can help me out with this.
This is what i have in mind:
Core tables
Category - Id int (PK) - Category varchar(100)
Question - Id int (PK) - Question varchar(200) - Category_Id int (FK- reference to Category.Id)
Answer - Id int (PK) - Answer - Question_Id int (FK- reference to Question.Id)
authorisation tables
UserType - Id int (PK) - UserType varchar(100)
User - Id int (PK) - Name varchar(100) - SID varchar(100) - UserType_Id int (FK - reference to UserType.Id)
Some questions: 1 i need a mechanism for the core tables to identify which records belongs to a certain user. So i think i need to add a foreign key to these tables. is this the right way to do this?
2 do i miss something important things for this design?
I am writing an asp application that uses sql server 2000 as the back-end db and I want to create a discussion board. I need to know what data type I should use to accomodate a large entry by the user and anything else I should be aware of.
There is now a NEW SQL discussion Group currently being tested; http://www.baysignia.com/discussions/discussions.xml Try it out and send comments to faq@baysignia.com
CREATE TABLE [dbo].[Audit] ( [id] [int] identity (1,1) NOT NULL , [col1] [char] (10) NOT NULL , [col2] [char] (15) NOT NULL , [col3] [int] NOT NULL , [col4] [varchar] (50) NOT NULL , [col5] [datetime] NOT NULL , [col6] [varchar] (4000) NULL , [col7] [char] (3) NULL ) GO
This table grows to about 40 million rows during the course of the month. The table has a clustered index on the id field and a non clustered index on the col2 and col3. The id column is not used in queries. At the moment we run weekly dbcc reindexes on all the indexes. We are running into a space issue on the reindex of the clustered index (copying the whole table out , ordering etc) and are considering dropping the index or changing to a non clustered index. (The DBCC utility that we have built will only rebuilt all the indexes or none at all.)
I feel this is not a good idea and know my reasons. I would like some input as to why this might prove a bad idea.
Will it increase page splitting? Will the table performance be impacted even if the queries are not specifically using the clustered index?
I'm creating a discussion forum for my website, using Sql Server 2000. I need to display 'Number of threads', 'Number of posts', 'Last post by' (username/id and date) for each forum, and 'Number of replies' and 'Last post by' (username/id and date) for each thread.Here are a couple of ideas I have come up with to solve this problem:1) Poll the database (using a stored procedure that returns the number I'm looking for) for each forum when I loop all the fourms. - I suspect this approach isn't optimal, since it creates more traffic to the database.2) Have fields in my Forum and ForumPost tables for 'Number of threads' and so on. Now, create triggers that updates these fields every time a post is made. - I guess this would be much more effective than the first apporach, since everything is done on the database server directly.Are there any other ways that are better? Please advice! Thanks a bunch for any help!
I'm sure this has been a topic in the past. I would like to build a fact sheet about each one and do a compairison. Are there any articles/links/resources out there that speak to this issue?
OK, I`ve been researching the use of the TEXT datatype all day and would like opinions on what I`ve found.
First, a little background. I have been tasked with writing an ASP application to handle the display of FAQs for a company`s products. I would like to store all info in a table much like
faqID int question TEXT answer TEXT
Simple enough, right? I then tried to create a stored procedure to add a new FAQ and all hell broke loose. ASP would not pass anything larger than 255 chars to the stored procedure.
I read in the "ADO and SQL Server Developer`s Guide" from Microsoft about using varchar datatypes of 255 chars (instead of TEXT) and chunking large text up to fit in these smaller datatypes. This seems like a lot of work.
I also read in "Inside SQL Server 6.5" that "The text datatype is sometimes awkward to work with. Many functions don`t operate against text, stored procedures are limited in what they can do with text, and some tools don`t deal with it well." (page 632). This statement concerns me greatly. How are stored procedures limited in dealing with TEXT? Do the standard SQL UPDATE and INSERT commands work or must READTEXT and UPDATETEXT be used instead?
I guess my question is, what is the best way to accomplish this? I have a feeling that others have had to do this before. Is SQL Server not meant to handle large textual objects? Is chunking the best way to go? Will version 7.0 handle this scenario better?
I would like to hear your thoughts on a philosophy I adhere to.
As a rule of thumb I've always preached that Unique Indexes are for linking tables and Primary Keys are used to ensure that records aren't duplicated. I’ve embraced this philosophy for a couple reasons, the main one being that I don’t have to create numerous foreign key fields in the foreign key table.
However I’ve done most of my programming in Access and am now in need of something more robust (SQL Server v7) and I’m wondering if I need to reconsider.
I do also have a how to question; that being is it possible to create a table join on a unique index in SQL Server v7 and if so how? I would like to have an Auto Number / Auto Incremented / Unique Identifier field in the Primary Key table that links to a numeric field in the Foreign Key table.
I have been doing relational database forever(or a long time) and have been intruduces to a team that uses a highly normalized database(propietary) to manage workflow.
We are capturing data in an AUDIT Trail EAV format.(500 million rows)
It is my task to build this into a data warehouse for reporting and I need to have with my team a relational database discussion. The relational database knowledge on this team is DB2 based, IDMS, and other past evolutions.
The common processes used are recieve a flat file and process this file sequentially using C# or VB doing lookups of other databse tables and writing out another flat file to be converted in XML for load to the propritary system.
My goal is to attempt to introduce new design concepts to my team and these are some talking points that I have come up with for a lunch and learn session.
can anyone else add to this list I don't want to get into a deep discussion about 3rd NF, Star Schemas vs Snowflake, etc.. I want to keep is informational and light to eliceit discussion and relat it back to older technologies.
some of the topics we can discuss are: Why the data warehouse Real-time tables what needs to stay in prod What is going to happen to reporting database Interaction between database on the same cluster/server Interaction between databases on different servers (linked servers not allowed) Set processing as opposed to cursor processing. Table types EAV Type1 Type2 Fact Dimensions Code
I'm a DB teacher in a University. Planning and developing SQL exercises for my students I found a "tricky" or extrange sql exercise.
The sql exercise using Northwind DB reads as follows: "Wich is the total amount of the freight that corresponds to all orders containing products of seafood category ?"
The first sql code that comes to my mind was:
select sum(freight) from orders o, "order details" d, products p, categories c where o.orderid=d.orderid and d.productid=p.productid and p.categoryid=c.categoryid and categoryname='seafood'
This is equivalent code constructed with Query constructor in Enterprise Manager: SELECT SUM(dbo.Orders.Freight) AS Expr1 FROM dbo.Orders INNER JOIN dbo.[Order Details] ON dbo.Orders.OrderID = dbo.[Order Details].OrderID INNER JOIN dbo.Products ON dbo.[Order Details].ProductID = dbo.Products.ProductID INNER JOIN dbo.Categories ON dbo.Products.CategoryID = dbo.Categories.CategoryID WHERE (dbo.Categories.CategoryName = N'seafood')
These 2 equivalent queries output 27722.9600 as result, but the correct result (total freight) is 23791.1400.
Did you also think (as I did) that the exposed code is correct?
Now I know the correct sql code that gives the corect answer and also the explanation why the code exposed fails, but before share it under your posible request, want to know your tech comments about this exercise in order to know if this particular exercise is "tricky" (make people to fail) or is my total fault and need to review my strategies of applying/constructing sql (and teaching).
I am about to prepare a paper concerning the field of real-time data mining. Real-time here means the process of incremental training of an existing model as soon as the data arrives.
There is a number of papers introducing algorithms for incremental association analysis, incremental clustering etc. Stream mining Ãs a field which is closely related to that. The main reason for the implementation of incremental algorithms is a) the large amount of data to be mined and b) the high rate of new data that is evolving every day.
Using classical batch mining algorithms, models that are outdated for some reason, would have to be re-trained, which could be very time consuming for billions of records. And once the training is completed, the training would have to be restarted once again because a bulk of new data has been arrived.
The question that I would like to discuss now is: For what real world applications would it be a meaningful or even essential to use real-time training of models?
Two main reasons could determine the answer to that question:
You just want to incorporate new data into existing models in order to increase the prediction accuracy of your model or Your underlying data is subject to more or less massive changes (also refered to as concept drift) and you want to adapt your mining model continuously to that reality.
I'm looking for some examples or ideas where one of these cases apply and it would be a good idea to have incremental mining algorithms involved.
I'm looking forward to inspiring some discussion on that issue.
in sql server 2000 ,in query analyzer :for example i have a emp table ,when i execute a query ,i have to get the emp table column names with its data types...for example :empno ---intempname -- varchar like this...what is the query to get the output like this
I am not using the SA account when I log in using windows authentication it seems to pick up my domain and active directory ID which is "usapp1dxd" it appears greyed out as well as the password on the login screen so you can't change it.
If you go into the management studio and right click on the server name in the left hand pane go to properties security the radio button for sql server and Windows authentication mode is selected.
For the begining thank for your reply. I started with a new job and I was hopeless when all normaly claimed things can't be done. I asked a lot of peoples who works on MSSQL(including SWYNK) for my problems(I know that I'm beginer on MSSQL) but they didn't knows any solution. And afterall I wrote that. For your ilustration, we develope information system for Invest company. Back is on MSSQL and client is in ASP.
>So... if you need a database you need to look at what needs to be done and >pick your DBMS that meets your requirements. --I didn't resolve on which DB we will develope:((
--No named cursors I mean that in general is using curosr names bootless. And you must deal with names... In interbase: DECLARE var1,2 INTEGER; FOR SELECT column,column2 FROM TABLE WHERE ... INTO :var1,:var2 BEGIN some code in loop; END
In mssql: DECLARE .. DECLARE crs CURSOR FOR SELECT ... OPEN crs FETCH NEXT FROM crs INTO @var1,@var2 WHILE (@@FETCH_RESULT = 0) BEGIN some code in loop FETCH NEXT FROM crs INTO @var1,@var2 END CLOSE crs DEALLOCATE crs -I thing that in mssql it isn't elegant. -problem with recursive procs -you must deal with names -If you can you may use named cursors in Interbase in fact
>-it don't know create resultset's from >> stored procs asynchronously when in sp is something else then only one >> select(problem if I want check access rigths to sp. for exapmle "...AS >> CheckPrivilege( ... ) SELECT.." This is confusing if user runs large >> query & he must wait until it creates the whole recordset...Armageddon goes >> first... Here is a solution, but it isn't very elegant and in some >> cases don't exists good solution > >I'm not sure if configuring SQL Server's "cursor threshold" parameter would solve this. "When >set to -1, all keysets are generated synchronously. If the cursor threshold is set to 0, all >keysets will be generated asynchronously."
-"cursor threshold" resolves it only in single select's and procedures with only one select's -It's unacceptable when your your user runs large queries
>>-max of nested procs is 16. >That is the limit in V6.5. In V7.0 the limit is raised to 32. -cool:)
>Union operator can't be used in a Create View statement in V6.5 according to the documentation. >Does Oracle support this? It appears that V7.0 of MSSQL will support UNION in a view. -cool:))
>-It don't have good exception > handling.(something like EXCEPTION, WHEN... (oracle, Intebase)) > >There are whole sections on error handling in SQL Books OnLine depending on >how you are >accessing SQL Server. Maybe it's as good as Oracle/Interbase... maybe not. I >don't seem to have a problem with it.
-Books are very feeble:( -but I thing that it can't work with EXCEPTION blocs. Good Exception handling is important tool for fast developing of robust apps
-I used example(s) from INTERBASE because I worked more on it, but in Oracle it's alike.
For a query like below ... How do i have to select only the latest revisions, if i need to filter last current revisions of each document ... where the revision could be either alphabetical or even numerical ...
Presently I get all revisions with the below query ... Note: csd_revi is the field of CSD table for revisions.
I've gotten everything to work -- almost! Here's the scoop... I'm using SQL Server 7.0 and ColdFusion (you don't need to know anything about ColdFusion). I'm trying to get SQL Server to publish/share/etc. a database with the network so ColdFusion (our website management/creation program) can access it and use it in a webpage. Well, so far, ColdFusion can detect and access the database (called "iami"), but it cannot find any tables in the database (particularily one called "phase1"). Can anyone help me -- the sooner the better -- ???
If you reply before 4pm today and offer truly helpful advice/input, I'll send you a reward/incentive from me personally (just give me your mailing address) just for helping out!!! :-)
Can anyone help me with handy scripts/stored procedures to capture blocking info on the server? We have SQL 7.0 w/sp3. Ray replied earlier saying that I can use so_who2!!, I think everyone knows that,please reply with some valuable info!!Please.. Thanks. Sonali.
I think jthis is a bug. I have a table created and populated on its own filegroup. I backup the db(all filegroups) and the trans log then I drop the one table. When I try to restore from my backups, it insists that I back up the trans log again. I do, then do the restore of both the filegroup and the trans log. The restore finishes, but my table is still not there and I can never get it back.
What should I have done? Is there anything that can be done other than restoring from backup? How does one know if the database is really recovering or is EM just joken? I can wait 2 hours before starting the restore
I was BCPing 12 million rows into a staging table. II used the '-b' option every 20K which I thought would do a commit and clear the log in batches. After the process EM appeared to show the transaction log as empty. Upon inspecting the Bcp output file I discovered the message that the BCP did not complete because syslogs was full. I could not do a truncate transaction log or a dump database. I tried to do a truncate transaction with no_log and it appeared to just hang. I stopped the SQL Server thinking I could dump the transaction log, but could not start the Sql Server again. I then stopped the NT Server because 'if all else fails'. The SQL Server started but the user database if marked as recovering.
Hi everyone I want to know if it's possible to do a for/while-loop so i can use INSERT
Look: I've this int [] test = new test[140]; But i need to insert for every value (140) a number so normally it would be : INSERT ... (case1, case2, case3 ...) value (test[1],test[2],test[3] ...) But isn't there a way to it with a loop? SOmething Like this ?
for( int i = 0 , i< 140, i++) { INSERT case[i] value test[i] }
Can it be possible to create trigger/procedurein following case2 server server A and server BA has db1db1 has tblAB has db2db2 has tblBcan it be possible to create trigger on server A.db1.tblA asinsert/update that trigger add/update record in server B.db2.tb2thnks-----------Hitendra
How do I indicate a post answered my question? While reading a reply to your question you will notice a button with this Icon: Clicking on it will mark the post as the answer to your question
I have set up a data driven subscription for a report, and against "Reply-To" I have selected "Specify a Static Value". I have then I have entered a valid email address.
When the report is received however, and the recipient clicks reply, the reply address is filled in as the Email Address from which the report was sent.
Is this a bug with the "Reply-To" functionality? Or do I have to do this a different way?
Is there a way to do this? When I have a further question in the same thread, my only option is to reply to the one who answered the last question. How do I put forward that question to everyone? Do I have to start a new thread with the same subject?
I'm trying to import data into an SQL Server (7.0) and I'm wondering which Source (Microsoft Data Link, Microsoft ODBC Driver for Oracle, Microsoft ODBC Driver for SQL Server, etc.) -- I THINK we would use the SQL Server driver but I'm not sure... to use AND WHERE TO GO FROM THERE? So far, I get seem to get things to work in my favor. I appreciate any help :) The data I'm trying to import is from Microsoft Excell. If there is anything else you need to know, please email me at iami@iami.org Please provide email/forum-based technical support.