Attribute 'importance' Factor
Jan 31, 2008
Is there a way to explicitly assign 'weights' or 'importance' factors to attributes and have that to be considered by the association rules and decision trees algorithms during training? I would like to do so without preprocessing the data (In any case, I can't think on a way to assign weight with preprocessing to boolean attributes like 'smoker')
thanks
View 3 Replies
ADVERTISEMENT
Aug 10, 2015
I want to insert attibute att1 in field F1 . Value for this attribute is content of another field in this table (F2).
My query :
update MyTable
set F1.modify
('insert attribute att1 {sql:column("F2")} into (/ROOT/Node1)[1]')
Where F1 Is not null
But I get this error :
XML well-formedness check: Duplicate attribute 'att1'. Rewrite your XQuery so it returns well-formed XML.
How do I check the douplicate attribute ?
View 0 Replies
View Related
May 16, 2008
HI,
I had to change the key columns of a dimension attribute to fix an error. I did this in BIDS. The change was from a single key column to a composite key column. Now I am getting these error when I process the cube:
Measure group attribute key column x does not match source attribute ..
I looked at the cube XMLA definition under mesaure groups and it still shows a single key column with inherited binding. However, the BIDS does not give me an option correct this in any way. I have had to do this once before and the only option seems to be removing the dimension from the cube and add it back in. But that is very error prone since I lose any specific settings at the cube dimension level not to mention aggregations no longer include the dimension, etc.
Not seeing an alternative, I went through each measure group (I have 7) and changed the key columns manually in the XMLA and saved the cube. This worked, but I don't understand why BIDS automatically doesn't do it.
Is this a flaw in the BIDS or I should be missing something.
thanks
MJ
View 3 Replies
View Related
Mar 12, 2001
Hi,
I am using Sql 7.0 with sp2. I just started as a sql dba. I have a question here, What is the importance of SID's ? When we are mapping to sql logins and user_id 's how we have to give importance regarding SID's.
Pls suggest me a good article or some suggestions...
Thanks!
View 1 Replies
View Related
Apr 27, 2007
hi i have an association rule mining model and i want to order the output by the importance here is the select statment:
SELECT
[RelatedOrder].[Order Line]
From
[RelatedOrder]
NATURAL PREDICTION JOIN
(SELECT (SELECT 888 AS [Product ID]) AS [Order Line]) AS t
thanx
View 3 Replies
View Related
Sep 30, 2004
Hi all-- there is a file in C:Program FilesMicrosoft SQL ServerMSSQLBinn direcetory called sqlctr.h which contains a lot of counter parametres..could any one tell me having its importance and can we change any of parametres to gain performance..
Thanks in advance..
// This file is generated by the description file processor.
// Please do not edit.
#defineBUFMGR_OBJECT0
#define BUF_RESERVED_PAGE_COUNT2
#define BUF_CHECKPOINT_WRITES4
#define BUF_AWE_LOOKUP_MAPS6
#define BUF_BLOCK_WRITES8
#define BUF_COMMITTED_PAGE_COUNT10
#define BUF_AWE_UNMAP_CALLS12
#define BUF_TARGET_PAGE_COUNT14
#define BUF_AWE_UNMAP_PAGES16
#define BUF_CACHE_RATIO_BASE18
#define BUF_FREELIST_STALLS20
#define BUF_HASHED_PAGE_COUNT22
#define BUF_LIFE_EXPECTANCY24
#define BUF_CACHE_HIT_RATIO26
#define BUF_AWE_WRITE_MAPS28
#define BUF_PAGE_REQUESTS30
#define BUF_STOLEN_PAGE_COUNT32
#define BUF_BLOCK_READS34
#define BUF_NUM_FREE_BUFFERS36
#define BUF_LAZY_WRITES38
#define BUF_READAHEAD_PAGES40
#define BUF_AWE_STOLEN_MAPS42
#define BUF_PROCCACHE_SIZE44
#defineBUFPART_OBJECT46
#define BUFPART_NUM_FREE_BUFFERS48
#define BUFPART_FREE_BUFFERS_USED50
#define BUFPART_FREE_BUFFERS_EMPTY52
#defineGENERAL_OBJECT54
#define GO_LOGINS56
#define GO_LOGOUTS58
#define GO_USER_CONNECTIONS60
#defineLOCKS_OBJECT62
#define LCK_TOTAL_WAITTIME64
#define LCK_NUM_WAITS66
#define LCK_AVERAGE_WAITTIME_BASE68
#define LCK_NUM_DEADLOCKS70
#define LCK_NUM_TIMEOUTS72
#define LCK_NUM_REQUESTS74
#define LCK_AVERAGE_WAITTIME76
#defineDBMGR_OBJECT78
#define DB_REPLTRANS80
#define DB_DBCC_SCANRATE82
#define DB_REPLCOUNT84
#define DB_LOG_SIZE86
#define DB_LOG_TRUNCS88
#define DB_LOG_USED_PERCENT90
#define DB_LOG_SHRINKS92
#define DB_BULK_KILOBYTES94
#define DB_FLUSH_WAIT_TIME96
#define DB_ACT_XTRAN98
#define DB_LOGCACHE_READS100
#define DB_FLUSH_WAITS102
#define DB_BCK_DB_THROUGHPUT104
#define DB_DBCC_MOVERATE106
#define DB_LOG_GROWTHS108
#define DB_TOTAL_XTRAN110
#define DB_LOGCACHE_BASE112
#define DB_BYTES_FLUSHED114
#define DB_LOG_USED116
#define DB_LOGCACHE_RATIO118
#define DB_DATA_SIZE120
#define DB_BULK_ROWS122
#define DB_FLUSHES124
#defineLATCH_OBJECT126
#define LATCH_TOTAL_WAIT_NP128
#define LATCH_WAITS_NP130
#define LATCH_AVG_WAIT_NP132
#define LATCH_AVG_WAIT_BASE134
#defineACCESS_METHODS_OBJECT136
#define AM_EXTENTS_ALLOCATED138
#define AM_WORKTABLES_CREATED140
#define AM_GHOSTED_SKIPS142
#define AM_FULL_SCAN144
#define AM_PAGES_ALLOCATED146
#define AM_PAGE_SPLITS148
#define AM_SINGLE_PAGE_ALLOCS150
#define AM_EXTENTS_DEALLOCATED152
#define AM_PROBE_SCAN154
#define AM_FREESPACE_PAGES156
#define AM_WORKTABLES_FROM_CACHE_BASE158
#define AM_LOCKESCALATIONS160
#define AM_PAGE_DEALLOCS162
#define AM_WORKTABLES_FROM_CACHE164
#define AM_INDEX_SEARCHES166
#define AM_FREESPACE_SCANS168
#define AM_FORWARDED_RECS170
#define AM_WORKFILES_CREATED172
#define AM_SCAN_REPOSITION174
#define AM_RANGE_SCAN176
#defineSQL_OBJECT178
#define SQL_AUTOPARAM_REQ180
#define SQL_BATCH_REQ182
#define SQL_RECOMPILES184
#define SQL_AUTOPARAM_UNSAFE186
#define SQL_COMPILES188
#define SQL_AUTOPARAM_FAIL190
#define SQL_AUTOPARAM_SAFE192
#defineCACHE_OBJECT194
#define CACHE_USE_COUNT196
#define CACHE_HIT_RATIO_BASE198
#define CACHE_OBJECT_COUNT200
#define CACHE_HIT_RATIO202
#define CACHE_PGS_IN_USE204
#defineMEMORY_OBJECT206
#define MEMORY_MEMGRANT_MAXIMUM208
#define MEMORY_CONNECTION_MEMORY210
#define MEMORY_MEMGRANT_WAITERS212
#define MEMORY_MEMGRANT_OUTSTANDING214
#define MEMORY_SQL_CACHE_MEMORY216
#define MEMORY_OPTIMIZER_MEMORY218
#define MEMORY_LOCKS220
#define MEMORY_SERVER_MEMORY222
#define MEMORY_LOCKOWNERS_ALLOCATED224
#define MEMORY_LOCK_MEMORY226
#define MEMORY_LOCKS_ALLOCATED228
#define MEMORY_SERVER_MEMORY_TARGET230
#define MEMORY_LOCKOWNERS232
#define MEMORY_MEMGRANT_ACQUIRES234
#defineUSER_QUERY_OBJECT236
#define QUERY_INSTANCE238
#defineREPLICATION_AGENT_OBJECT240
#define RUNNING_INSTANCE242
#defineMERGE_AGENT_OBJECT244
#define MERGE_CONFLICTS_INSTANCE246
#define UPLOAD_INSTANCE248
#define DOWNLOAD_INSTANCE250
#defineLOGREADER_AGENT_OBJECT252
#define LOGREADER_LATENCY_INSTANCE254
#define LOGREADER_TRANSACTIONS_INSTANCE256
#define LOGREADER_COMMANDS_INSTANCE258
#defineDISTRIBUTION_AGENT_OBJECT260
#define DISTRIBUTION_TRANS_INSTANCE262
#define DISTRIBUTION_LATENCY_INSTANCE264
#define DISTRIBUTION_COMMANDS_INSTANCE266
#defineSNAPSHOT_AGENT_OBJECT268
#define SNAPSHOT_TRANSACTIONS_BCPED270
#define SNAPSHOT_COMMANDS_BCPED272
#defineBACKUP_DEV_OBJECT274
#define DB_BCK_DEV_THROUGHPUT276
View 3 Replies
View Related
Jul 17, 2007
Hello Experts at Microsoft.
I am thinking of an easy way to explain importance to Marketers without going into the math. This is what i came up with so far. Does this sound correct to you guys?
Reasoning:
IMPORTANCE = Log(Improvement)
Improvement=P(X&Y)/(P(x)*P(y))
Improvement= (Probability 2 products are sold together)/(random chance 2 products are sold together)
If the (Probability 2 products are sold together) = (random chance 2 products are sold together) then Improvement=1. The log(1) = 0
IMPORTANCE SCORE
-2 to -1 10 to 100 times less likely than random chance
-1 to 0 0 to 10 times less likely than random chance
0 to 1 0 to 10 times more likely than random chance
1 to 2 10 to 100 times more likely than random chance
2 to 3 100 to 1000 times more likely than random chance
3 to 4 1000 to 10000 times more likely than random chance
4 to 5 10000 to 100000 times more likely than random chance
5 to 6 100000 to 1000000 times more likely than random chance
6 to 7 1000000 to 10000000 times more likely than random chance
View 1 Replies
View Related
Feb 14, 2008
I understand Mr. MacLennan's explanation provided at http://forums.microsoft.com/MSDN/ShowPost.aspx?PostID=282651&SiteID=1 and appreciate the time he took to explain how importance works. However, like the user with username "sang", I also ran the data in BI 2005 and got the same results listed by the aforementioned user. I did this using the following data:
donut
muffin
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
n
y
n
y
n
y
n
y
n
y
etc.
The rule muffin -> donut has an importance of -0.105302438, which is not the same as Mr. MacLennan's results. I tried switching the roles of a and b in a -> b and using different bases on the logarithms. I don't get the result of -0.105302438 with any of these. I also tried to calculate importance with a small data set I have and can't get the results using Mr. MacLennan's explanation with that data set either. Any thoughts on the descrepancy?
View 5 Replies
View Related
Feb 21, 2008
WHY DO WE USE TRIGGERS IN SQL SERVER2005.
WAT IS ITS IMPORTANCE.
AND SOME SAMPLES
PLEASE GIVE ME SOLUTIONS
View 1 Replies
View Related
Apr 12, 2007
hi,
i have a exercise using association datamining
my database have 350 records,
i use 90 records for datamining and it release some rules which i choose on top of mSOLAP_NODE_SCORE,
but when i use select statement to check my result i have 1 records, the same as my result, and 5 records not true;
for example:
rules A=a,B=b-> C=c
select * from <my_table> where A='a' and B='b' and C='c'; ==>1 record return
select * from <my_table> where A='a' and B='b' and C<>'c'; ==>5 records return
C with 3 values c1,c2,c
with the second statement C includes 2 c1 and 3 c2
i don't understand how they work.
i want to choose some best rules can present my database.
how can i choose importance and probability to get best rules.
with database have 90 records and a database have 350 records which values i should use for minimum_probability, Minimum_Support, Minimum_importance...
when i choose rules i should choose on importance or probability.
thanks for your help
View 4 Replies
View Related
Mar 6, 2006
Can anyone tell me, how the Business Ã?ntelligence Studio calculates the importance of a rule. I can't find the formula. I know some formulas, but the result in SQL Server is completly different.
Thanks!
View 12 Replies
View Related
Oct 19, 2005
Those of you who have installed SQL Server 2005 may have noticed that the installation creates several new Windows groups on the server. Do not underestimate the importance of these groups.
View 3 Replies
View Related
Nov 5, 2003
Does anybody know how to send an email using xp_sendmail sp with HIGH importance setting for the message?
Thanks,
Dim
View 5 Replies
View Related
Dec 30, 2004
I try to search my data and sort the result by importance.
I'm using a MS Access database and my data (table1) looks like this:
Code:
ID NAME TEXT
1 Apples Good red apples
2 Bananas Fine yellow bananas
3 Yellow apples Great yellow apples
I want to search the data and get a result where the column "NAME" is more important than "TEXT". My SQL looks like this:
Code:
SELECT id,name,text,1 AS searchorder FROM table1 WHERE name LIKE '*yellow*'
UNION
SELECT id,name,text,2 AS searchorder FROM table1 WHERE text LIKE '*yellow*'
ORDER BY searchorder
The output is this:
Code:
ID NAME TEXT SEARCHORDER
3 Yellow apples Great yellow apples 1
2 Bananas Fine yellow bananas 2
3 Yellow apples Great yellow apples 2
So far so good - the order by importance works - but I do not get unique columns because of the searchorder column.
Can I fix my SQL so I get unique columns where the last line of "Yellow apples" does not appear or am I lost in space?
Best regards,
Peter from Denmark
View 2 Replies
View Related
Sep 22, 2006
During testing a package repetatively that deletes/inserts into several tables, over the course of several days, my package, which took 45 minutes to load 1700 XML files, began to take over 6 hours. Turns out it was an I/O bottleneck, and the Avg Disk Queue Length was around 200 and I was incurring many PAGEIOLATCH_EX. My devl machine uses a single local disk, no raid, so I had no options there, but I ran the maintenance wizard to recreate indexes/statistics and defraged the hard drive, and regained my original 45 minutes time. I guess I'll have to put a maintenance plan together to do this nightly.
-Kory
View 1 Replies
View Related
Apr 3, 2001
Currently we have tables (in sql 6,5), many of them do not have primary keys.
While I was trying to re-index (re-org), many of them got an error:
"fillfactor 204 is not a valid percentage; fillfactor must be between 1 and 100."
(many tables' fillfactor exceed 100 or more...)
How can I fix them so I can upgrade to sql 7 ?
Thank you for your help.
View 3 Replies
View Related
Aug 13, 2001
I am really confused about this whole fill factor thing. The way I understand it, is if you have a table whose data remains pretty much static, you should use a higher fill factor. Suppose you had a database where you had at most 150 transactions a day that changed the data, should the fill factor be left at the default(0) or increased? How do you determine how much to increase it? Is there a rule of thumb that suggests if you have x number of changes against a table, you should have a fill factor between y and z percent?
Please Help
Chris
View 2 Replies
View Related
Mar 12, 2001
Hi all,
While creating indexes for a table, I specified a fill factor of 70%. I then inserted a few hundred rows into the table. Is it possible to check to what percent the pages are full after the rows have been inserted?
Thanks in advance,
Praveena
View 1 Replies
View Related
Jul 16, 1999
You have a db with 50,000 records and you want to add 100,000 more. What should the right fill factor be? Is there a way to "calculate" a fill factor if you don't want to use default? Any help is appreciated. Thank you.
View 3 Replies
View Related
Jan 18, 2005
If fill factor is specified as 100 for a table what will be the impact of this?I want to know any updation or insertion will be possible or not?
View 3 Replies
View Related
Aug 27, 2007
Hellow, everyone"
I have a web online table that is inserted about 1500 record one day. Each night, a DST is running to pull all data to anther database. How to set fill factor on a one column index to get the best performance? Current fill factor is 80%.
Thanks
ZYT
View 7 Replies
View Related
May 20, 2008
How Index will help to optimize SP.
spatle
View 1 Replies
View Related
Jul 13, 2007
Hi experts, I would like to ask regarding FILL FACTOR. I observed that our system's loading is a bit slow, and some of the modules take 1 to 2 minutes loading. Maintenance activity is regularly executed based on the scheduled sets. Then I tried to checked the tables indexes/keys turns out that the FILL FACTOR is set to ZERO(0). I would like to know if the FILL FACTOR set to zero will be a factor for the system to slow down..????
Darren Bernabe Blanco
View 2 Replies
View Related
Jul 14, 2006
How do i find out how much fill facot has been spcified in given table?.
View 1 Replies
View Related
Apr 4, 2001
What are the criteria to decide the optimal(or close to optimal) fill factor?
Any input is appreciated.
View 2 Replies
View Related
Aug 9, 2001
You have 50,000 records in a database file and you know you want to add another 100,000 records in the next several weeks. What fill factor would you use to maximize performance?
A.0(default setting)
B.30
C.70
D.100
which one is correct? And how to calculate fill factor?
View 1 Replies
View Related
Sep 14, 1998
I have some non-clustered , non-unique indexes on a medium sized table (25,000 rows).
The fill factore is showing 248% on these indexes. I have tried setting the fill factor to various values 100% or less. The re build index seems to work, however the est min/avg/max size of the index all show appx 160Kb, whilst the actual size is in excess of 50Mb !!
We run a weekly rebuild of all indexes overnight, without any fill params and following this the fill on these indexes goes back to 248% !
I have also dropped and re-created the index with a fill of 100%, and it has still reverted to 248% following the weekly rebuild.
I have also looked at the server config, and the fill factor there shows a running value of 171% !!, although the current config is set to 0.
The server is stopped every evening, so no way should the config have a value of 171, especially since the max allowed value is 100.
Any advice/assistance would be gratefully received.
View 1 Replies
View Related
Jun 7, 2002
I know what fill factor is ... and know that I should set it high when I have static data tables (where the data rarely changes) and low when I expect to have page splits ...
but does anyone know what affect on performance this setting has ? I don't quite get what Books Online says about it.
Thanks
Carol
View 1 Replies
View Related
Aug 25, 2004
Hi ,
Can u plz tell me what is fill factor and what its role in defining the indexes. It is by default 0% and can be set upto 100%, but what it makes difference if i change the percentage? Where it is exactly impacts? If u know any links then plz forward me.
Thanks And Regards,
Shailesh
View 3 Replies
View Related
Oct 13, 2004
I am trying to set up the relationship (Primary Key and Forign Key) in several tables. I would like to find a way also be able to set 'Fill Factor = 90%' in the script. Here is the code that i have so far:
ALTER TABLE Req
ADD CONSTRAINT FK_Req_Bow FOREIGN KEY ( BowID )
REFERENCES Bow ( BowID )
Any help would be greatly appreciated!
J827
View 1 Replies
View Related
Feb 21, 2008
Hi there
Is there any such thing that you can find the current Fill Factor for each indexing? The only thing that you have an indication is by looking at DBCC SHOWCONTIF > Scan Density [Best Count:Actual Count].......: 100.00% [0:0] and if this value is not reaching 100% means may have an issue with fill factor ?!!?
Also, how about SQL2000?
Thanks
View 4 Replies
View Related
Jul 20, 2005
HiThere are a lot of articles about the fillfactor.I did change the fill factor and that did not work as intended.How do I get back to the default fill factorI am using sp_msforeachtable undocumented database procedureand when the indexes are rebuilt the fill factor that shows upin origfillfactor is the one I am trying to move away fromYour help will be appreciatedVince
View 4 Replies
View Related
Jul 31, 2007
Turn away pure key zealots
I have a clustered index that starts with an INT IDENTITY(1,1) column and therefore you can only add data to the end of the cluster. What I'm confused about is the relationship between this and the fill factor. In a normal fill factor scenario you'd be worried about inserts causing page splits but if you can only append to this cluster does this mean I should set the factor to 100% even if I'm expecting a large number of inserts? Basically I don't understand what happens when you run out of space on a page on a B-tree if it's based upon an ever increasing number.
View 3 Replies
View Related