亚洲欧美第一页_禁久久精品乱码_粉嫩av一区二区三区免费野_久草精品视频

? 歡迎來到蟲蟲下載站! | ?? 資源下載 ?? 資源專輯 ?? 關于我們
? 蟲蟲下載站

?? c4.5文檔說明.txt

?? C4.5文檔說明(數據類型
?? TXT
?? 第 1 頁 / 共 4 頁
字號:

In the hypothyroid example, using a sample of 60% would cause a classifier to be constructed from a randomly-selected 1663 of the 2772 cases in hypothyroid.data, then tested on the remaining 1109 cases. 

By default, the random sample changes every time that a classifier is constructed, so that successive runs of See5 with sampling will usually produce different results. This re-sampling can be avoided by selecting the Lock sample option that uses the current sample for constructing subsequent classifiers. If this option is selected, the sample will change only when another application is loaded, the sample percentage is altered, the option is unselected, or See5 is restarted. 


Cross-validation trials
As we saw earlier, the performance of a classifier on the training cases from which it was constructed gives a poor estimate of its accuracy on new cases. The true predictive accuracy of the classifier can be estimated by sampling, as above, or by using a separate test file; either way, the classifier is evaluated on cases that were not used to build it. However, this estimate can be unreliable unless the numbers of cases used to build and evaluate the classifier are both large. If the cases in hypothyroid.data and hypothyroid.test were to be shuffled and divided into a new 2772-case training set and a 1000-case test set, See5 might construct a different classifier with a lower or higher error rate on the test cases. 

One way to get a more reliable estimate of predictive accuracy is by f-fold cross-validation. The cases in the data file are divided into f blocks of roughly the same size and class distribution. For each block in turn, a classifier is constructed from the cases in the remaining blocks and tested on the cases in the hold-out block. In this way, each case is used just once as a test case. The error rate of a classifier produced from all the cases is estimated as the ratio of the total number of errors on the hold-out cases to the total number of cases. 

The Cross-validation option with f folds runs such a f-fold cross-validation. Suppose now that we select the Cross-validation option with 10 folds together with the Rulesets option. After giving details of the individual rulesets, the output shows a summary like this: 


	Fold	        Rules     
	----	  ----------------
		    No      Errors
	
	   0	     7        0.8%
	   1	     7        0.3%
	   2	     7        0.5%
	   3	     7        0.3%
	   4	     8        0.8%
	   5	     7        0.5%
	   6	     7        0.3%
	   7	     6        0.5%
	   8	     7        0.5%
	   9	     7        0.8%
	
	  Mean	   7.0        0.5%
	  SE	   0.1        0.1%

This estimates the error rate of the rulesets produced from the 2772 cases in hypothyroid.data at 0.5%. The SE figures (the standard errors of the means) provide an estimate of the variability of these results. 

The cross-validation procedure can be repeated for different random partitions of the cases into blocks. The average error rate from these distinct cross-validations is then an even more reliable estimate of the error rate of the single classifier produced from all the cases. 

Since every cross-validation fold uses only part of the application's data, running a cross-validation does not cause a classifier to be saved. To save a classifier for later use, simply run See5 without employing cross-validation. 


Differential misclassification costs
Up to this point, all errors have been treated as equal -- we have simply counted the number of errors made by a classifier to summarize its performance. Let us now turn to the situation in which the `cost' associated with a classification error depends on the predicted and true class of the misclassified case. 

See5 allows costs to be assigned to any combination of predicted and true class via entries in the optional file filestem.costs. Each entry has the form 


	predicted class, true class: cost

where cost is any non-negative value. The file may contain any number of entries; if a particular combination is not specified explicitly, its cost is taken to be 0 if the predicted class is correct and 1 otherwise. 

To illustrate the idea, suppose that it was a much more serious error to classify a hypothyroid patient as negative than the converse. A hypothetical costs file hypothyroid.costs might look like this: 


	negative, primary: 5
	negative, secondary: 5
	negative, compensated: 5

This specifies that the cost of misclassifying any primary, secondary, or compensated patient as negative is 5 units. Since they are not given explicitly, all other errors have cost 1 unit. In other words, the first kind of error is five times more costly. 

A costs file is automatically read by See5 unless the system is told to ignore it. The output from the system using default options now looks like this: 


	See5 [Release 1.20a]	Wed Sep  1 11:04:52 2004

	Class specified by attribute `diagnosis'
	
	Read 2772 cases (24 attributes) from hypothyroid.data
	Read misclassification costs from hypothyroid.costs
	
	Decision tree:
	
	TSH <= 6:
	:...TT4 > 54: negative (2444.3)
	:   TT4 <= 54:
	:   :...referral source in {WEST,SVHC,SVHD}: negative (0)
	:       referral source = STMW: negative (1)
	:       referral source = SVI: negative (18)
	:       referral source = other:
	:       :...T4U > 0.88: secondary (3.5/2.1)
	:           T4U <= 0.88:
	:           :...query hypothyroid = f: negative (3.6)
	:               query hypothyroid = t: secondary (1.7/1.1)
	TSH > 6:
	:...FTI <= 65:
	    :...TT4 <= 63: primary (59.4/8)
	    :   TT4 > 63:
	    :   :...T4U <= 1.1: compensated (2.9/1.3)
	    :       T4U > 1.1:
	    :       :...TT4 <= 90: primary (8.8/1.7)
	    :           TT4 > 90: compensated (1.3/0.2)
	    FTI > 65:
	    :...on thyroxine = t: negative (37.7)
	        on thyroxine = f:
	        :...thyroid surgery = t: negative (6.8)
	            thyroid surgery = f:
	            :...TT4 <= 61:
	                :...TT4 <= 37: primary (2.5/0.2)
	                :   TT4 > 37: compensated (3.4/0.4)
	                TT4 > 61:
	                :...age > 8:
	                    :...TT4 <= 144: compensated (163.7/22.7)
	                    :   TT4 > 144:
	                    :   :...TT4 <= 153: compensated (2.3/0.3)
	                    :       TT4 > 153: negative (6/0.1)
	                    age <= 8:
	                    :...TSH > 29: primary (0.7)
	                        TSH <= 29:
	                        :...referral source in {WEST,SVHC,SVI,
	                            :                   SVHD}: compensated (0)
	                            referral source = other: compensated (2.8)
	                            referral source = STMW:
	                            :...age <= 1: compensated (1)
	                                age > 1: primary (0.7)
	
	
	Evaluation on training data (2772 cases):
	
		       Decision Tree       
		  -----------------------  
		  Size      Errors   Cost  
	
		    21   11( 0.4%)   0.00   <<
	
	
		   (a)   (b)   (c)   (d)    <-classified as
		  ----  ----  ----  ----
		    62     1                (a): class primary
		         154                (b): class compensated
		                 2          (c): class secondary
		     6     2     2  2543    (d): class negative
	
	
	Evaluation on test data (1000 cases):
	
		       Decision Tree       
		  -----------------------  
		  Size      Errors   Cost  
	
		    21    8( 0.8%)   0.01   <<
	
	
		   (a)   (b)   (c)   (d)    <-classified as
		  ----  ----  ----  ----
		    31     1                (a): class primary
		     1    39                (b): class compensated
		                            (c): class secondary
		     2     2     2   922    (d): class negative
	
	
	Time: 0.0 secs

This new decision tree has a higher error rate than the first decision tree for both the training and test cases, and might therefore appear entirely inferior to it. The real difference comes when we compare the total cost of misclassified training cases for the two trees. The first decision tree, which was derived without reference to the differential costs, has a total cost of 19 (4x1 + 3x5) for the misclassified training cases in hypothyroid.data. The corresponding value for the new tree is 11 (11x1). That is, the total misclassification cost over the training cases is lower than that of the original tree. The total misclassification cost on the test data is 8 (3x1 + 1x5) for the original tree and also 8 (8x1) for the new tree. 

Using Classifiers
Once a classifier has been constructed, an interactive interpreter can be used to predict the classes to which new cases belong. The Use Classifier button invokes the interpreter, using the most recent classifier for the current application, and prompts for information about the case to be classified. Since the values of all attributes may not be needed, the attribute values requested will depend on the case itself. When all the relevant information has been entered, the most likely class (or classes) are shown, each with a confidence value. For example, this is the result of analyzing a case using the first decision tree above: 

 

Classifiers can also be used in batch mode. The sample application provided in the public source code reads cases from a cases file and shows the predicted class and the confidence for each. 


Cross-Referencing Classifiers and Data
See5 incorporates a unique facility that links data and the relevant sections of (possibly boosted) classifiers. We will illustrate this facility using the first decision tree for the hypothyroid application and the cases in hypothyroid.data from which it was constructed. 

The Cross-Reference button brings up a window showing the most recent classifier for the current application and how it relates to the cases in the data, test or cases file. (If more than one of these is present, a menu will prompt you to select the file.) 

The window is divided into two panes, with the classifier on the left and a list of cases on the right. The Reset button can be used at any time to restore the window to this initial state. 

Each case has a [?] tag (that is red if the case is misclassified), an identifying number or label, and the class predicted for the case (also red when incorrect). Clicking on the tag [?] in front of a case number or label displays that case: 

 

The values of label attributes and attributes excluded or ignored are displayed in a lighter tone to indicate that they play no part in classifying the case. 

Clicking on a case's label or number shows the part(s) of the classifier(s) relevant to that case. For instance, clicking on case 3169 shows the leaf to which this case is mapped: 

 

If a case has missing values for one or more attributes, if it is covered by several rules, or if boosted classifiers are used, more than one leaf or rule may be relevant to a case. In such situations, all relevant classifier parts are shown. 

Click on any leaf or rule, and all the cases that map to the leaf or rule are shown. For instance, clicking on Reset and then the leaf indicated shows all cases that are covered by that leaf: 

 

This last pane may be puzzling for two reasons: 

The case pane shows nine cases but the count shown at the leaf is 3.8. This happens because some of these nine cases have unknown values for the attributes tested on the path to this leaf (TSH, FTI, thyroid surgery, TT4, referral source). Cases like this are split into partial cases associated with each outcome of the test. 
This leaf predicts class primary, but some cases belonging to other classes are not highlighted in red to indicate an error. As noted above, parts of a case split as a result of unknown attribute values can be misclassified and yet, when the votes from all the parts are aggregated, the correct class can still be chosen. Cases 3469, 3640, 2266, 311, and 3607 are classified correctly by the decision tree as a whole. 
The Save button preserves the details of the displayed classifier and case list as an ASCII file selected through a dialog box. 

Generating Classifiers in Batch Mode
The See5 distribution includes a program See5X that can be used to produce classifiers non-interactively. This console application resides in the same folder as See5 (usually C:\Program Files\See5 for single-computer licences or the See5 folder on your desktop for network licences) and is invoked from an MS-DOS Prompt window. The command to run the program is: 

	start See5X -f filestem parameters
where the parameters enable one or more options discussed above to be selected: 
-s  use the Subset option  
-r  use the Ruleset option  
-u bands  sort rules by their utility into bands  
-b  use the Boosting option with 10 trials  
-t trials  ditto with specified number of trials  
-w  winnow attributes before constructing a classifier  
-S x  use the Sampling option with x%  
-I seed  set the sampling seed value  
-X folds  carry out a cross-validation  
-g  turn off the global tree pruning stage  
-c CF  set the Pruning CF value  
-m cases  set the Minimum cases  
-p  use the Fuzzy thresholds option  
-e  ignore any costs file  

If desired, output from See5 can be diverted to a file in the usual way. 

As an example (for a single-computer licensee), typing the commands 

	cd "C:\Program Files\See5"
	start See5X -f Samples\anneal -r -b >save.txt

in a MS-DOS Prompt window will generate a boosted ruleset classifier for the anneal application in the Samples directory, leaving the output in file save.txt. 
Linking to Other Programs
The classifiers generated by See5 are retained in files filestem.tree (for decision trees) and filestem.rules (for rulesets). Free C source code is available to read these classifier files and to make predictions with them, enabling you to use See5 classifiers in other programs. As an example, the source includes a program to read cases from a cases file, and to show how each is classified by boosted or single trees or rulesets. 

Click here to download a zip archive containing the public source code. 

? RULEQUEST RESEARCH 2004  Last updated September 2004  



--------------------------------------------------------------------------------
home  products  download  evaluations  prices  purchase  contact us  

?? 快捷鍵說明

復制代碼 Ctrl + C
搜索代碼 Ctrl + F
全屏模式 F11
切換主題 Ctrl + Shift + D
顯示快捷鍵 ?
增大字號 Ctrl + =
減小字號 Ctrl + -
亚洲欧美第一页_禁久久精品乱码_粉嫩av一区二区三区免费野_久草精品视频
jlzzjlzz欧美大全| 麻豆国产精品视频| 亚洲午夜视频在线观看| 亚洲成人av在线电影| 免费观看日韩av| 高清成人免费视频| 色www精品视频在线观看| 91精品国产一区二区三区香蕉| 日韩美女在线视频| 亚洲欧洲美洲综合色网| 亚洲国产精品久久人人爱蜜臀| 国产在线视视频有精品| 97久久超碰国产精品| 91麻豆精品国产91久久久使用方法| 久久人人爽人人爽| 亚洲精品国产成人久久av盗摄| 日本欧美大码aⅴ在线播放| 成人一二三区视频| 欧美人妇做爰xxxⅹ性高电影| 久久久久88色偷偷免费| 亚洲高清免费一级二级三级| 国产成人在线观看| 欧美老年两性高潮| 国产精品丝袜一区| 奇米影视7777精品一区二区| 97精品视频在线观看自产线路二| 日韩欧美国产三级电影视频| 伊人色综合久久天天人手人婷| 美女视频免费一区| 一本一道久久a久久精品| 欧美精品一区二区在线观看| 亚洲自拍偷拍图区| 国产乱码精品一区二区三区av| 在线观看日韩一区| 国产精品乱码一区二区三区软件 | 欧美在线三级电影| 国产欧美日韩在线| 久久精品国产亚洲5555| 在线观看网站黄不卡| 国产日产欧美一区| 激情综合五月婷婷| 91精品欧美久久久久久动漫| 亚洲欧美国产毛片在线| 韩国毛片一区二区三区| 777精品伊人久久久久大香线蕉| 综合色天天鬼久久鬼色| 国产真实乱偷精品视频免| 欧美一级日韩免费不卡| 亚洲在线成人精品| 99麻豆久久久国产精品免费| 久久精品亚洲乱码伦伦中文| 蜜臀a∨国产成人精品| 欧美日韩亚洲综合| 亚洲美女在线一区| 99精品视频一区二区三区| 久久久久九九视频| av午夜一区麻豆| 精品捆绑美女sm三区| 秋霞午夜av一区二区三区| 欧美日韩精品一区二区三区| 亚洲人妖av一区二区| 成人av免费观看| 国产欧美1区2区3区| 国产精品亚洲视频| 国产性天天综合网| 国产suv精品一区二区6| 337p粉嫩大胆色噜噜噜噜亚洲| 免费日本视频一区| 91精品国产手机| 免费在线视频一区| 日韩欧美色综合网站| 免费观看成人av| 欧美一级在线观看| 狠狠色综合色综合网络| 欧美一区二区三区免费在线看| 全国精品久久少妇| 欧美va亚洲va国产综合| 看电影不卡的网站| 精品粉嫩超白一线天av| 国内国产精品久久| 2020国产精品久久精品美国| 国产在线播放一区三区四| 久久婷婷色综合| 高清国产一区二区| 亚洲人成伊人成综合网小说| 欧美亚洲国产一区二区三区 | 欧美性猛交xxxxxx富婆| 亚洲电影在线播放| 91精品国产91久久久久久最新毛片| 日韩精品每日更新| 欧美成人激情免费网| 国产一区 二区| 国产精品久久毛片| 日本韩国欧美国产| 天堂成人国产精品一区| 欧美tickling挠脚心丨vk| 国产精品中文字幕一区二区三区| 日本一区二区三区四区| 色国产综合视频| 日韩精品亚洲专区| 久久综合国产精品| 99久久综合99久久综合网站| 亚洲日本在线天堂| 欧美军同video69gay| 国产在线一区二区综合免费视频| 国产精品区一区二区三区| 欧美性受xxxx| 国产精品一区二区久激情瑜伽| 中文字幕综合网| 欧美乱妇20p| 国产91丝袜在线播放| 亚洲一区二区三区自拍| 欧美成人午夜电影| 99re亚洲国产精品| 日日欢夜夜爽一区| 欧美国产日韩精品免费观看| 91黄色激情网站| 免费在线看一区| 日韩理论片在线| 欧美一区二区三区免费视频| 成人午夜视频网站| 丝袜美腿亚洲色图| 中文字幕av一区 二区| 欧美无人高清视频在线观看| 九一九一国产精品| 亚洲欧美偷拍另类a∨色屁股| 91精品国产aⅴ一区二区| 成人高清伦理免费影院在线观看| 午夜视频一区二区三区| 国产农村妇女精品| 欧美日韩视频在线一区二区| 国产乱码精品一区二区三| 亚洲国产精品一区二区尤物区| 精品处破学生在线二十三| 精品久久人人做人人爰| 在线欧美小视频| 国产盗摄精品一区二区三区在线 | 正在播放亚洲一区| 成人黄色777网| 久久精品国产亚洲高清剧情介绍| 亚洲欧美视频在线观看| 26uuu成人网一区二区三区| 在线亚洲人成电影网站色www| 极品少妇xxxx精品少妇| 亚洲一区二区三区不卡国产欧美| 欧美精品一区二区在线播放| 欧美日韩不卡一区二区| 成人a区在线观看| 国产一区二区三区在线观看精品| 亚洲成av人综合在线观看| 亚洲欧洲国产专区| 久久九九久精品国产免费直播| 91精品欧美福利在线观看| 色94色欧美sute亚洲13| 99久久久精品| 国产99久久久精品| 另类小说欧美激情| 日韩在线一区二区三区| 亚洲国产日韩一区二区| 亚洲婷婷综合色高清在线| 久久久精品综合| 亚洲精品在线一区二区| 91精品国产福利在线观看| 欧美日韩精品电影| 91国偷自产一区二区三区观看 | 亚洲激情校园春色| 成人欧美一区二区三区黑人麻豆 | 国产a视频精品免费观看| 蜜臀a∨国产成人精品| 日韩一区精品视频| 午夜精品在线看| 亚洲一区二区欧美日韩| 亚洲精品久久嫩草网站秘色| 国产精品久久久久精k8| 欧美国产一区二区在线观看| 2021中文字幕一区亚洲| 精品对白一区国产伦| 精品福利在线导航| 精品国产成人系列| 欧美本精品男人aⅴ天堂| 日韩一区二区三区免费看| 91麻豆精品久久久久蜜臀| 欧美丰满美乳xxx高潮www| 欧美精选午夜久久久乱码6080| 欧美日韩专区在线| 欧美精品成人一区二区三区四区| 欧美手机在线视频| 欧美伦理影视网| 7777精品伊人久久久大香线蕉完整版| 欧美色精品在线视频| 欧美午夜电影一区| 欧美日韩电影一区| 在线成人小视频| 欧美一区2区视频在线观看| 日韩一级片网站| 精品国产自在久精品国产| 2020国产精品| 国产精品免费视频观看| 国产精品久久久久影院色老大| 青青草伊人久久|