全部版块 我的主页
论坛 计量经济学与统计论坛 五区 计量经济学与统计软件
5753 5
2005-01-18

In XLMinerTM, select Classification --> Discriminant Analysis. In the dialog box that comes up, you can specify the data to be used, the input variables and the output variable.

Variables: This box lists all the variables present in the dataset. If the "First row contains headers" box is checked, the header row above the data is used to identify variable names.

Variables in input data: Select one or more variables as independent variables from the Variables box by clicking on the corresponding selection button. These variables constitute the predictor variables.

Output Variable: Select one variable as the dependent variable from the Variables box by clicking on the corresponding selection button. This is the variable being classified.

Specify "Success" class : In logistic regression the output variable has catagorical values. Eg. Let us enter a value "1" here. Then, if in a record the output variable attains a value of 1 in the training data, that is taken as success.

Specify initial cutoff probability value for success : Enter the desired value here, say 0.5. Then the class is taken to be a success if the probability is greater than this value.

Click Next and the following dialog box appears:

Calculate according to relative occurrences: The discriminant analysis procedure incorporates prior assumptions about how frequently the different classes occur. If this option is checked, it will be assumed that the probability of encountering a particular class in the large data set is the same as the frequency with which it occurs in the training data.

Use equal prior probabilities: If this option is checked, it will be assumed that all classes occur with equal probability.

Click Next and, in the following dialog box, choose the required outputs:

Canonical variate loadings: XLMinerTM produces the canonical variates for the data which is based on an orthogonal representation of the original variates. This has the effect of choosing a representation which maximizes the distance between the different groups. For a k class problem there are k-1 Canonical variates. Very often only a subset (say g) of the canonical variates is sufficient to discriminate between the classes.

Canonical Scores: The values of the variables X1, X2, ...Xg for the ith observation are known as the canonical scores for that observation. The purpose of the canonical score is to make separation between the classes as large as possible. Thus when the observations are plotted with the canonical scores as the coordinates, the observations belonging to same class are grouped together.

Score training / validation data: Check appropriate options to show the scores of training and validation data.

Score Test/New Data: Select the appropriate option for applying the model to test data and / or new data as required. See the Example of Discriminant Analysis for detailed instructions on new data.

Score New data in database : See the Example of Discriminant Analysis for detailed instructions on this.

Click Finish and the output will be displayed as per the inputs given in the dialogs above.

See also

二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

全部回复
2005-1-18 12:34:00

[推荐]Using Neural Network Classification in XLMiner

Using Neural Network Classification in XLMinerTM:

In XLMinerTM, select Classification -> Neural Network. This brings up the following dialog box, where you need to specify the data range to be processed, the input variables and the output variable.

Variables: This box lists all the variables present in the dataset. If the "First row contains headers" box is checked, the header row above the data is used to identify variable names.

Variables in input data: Select one or more variables as independent variables from the Variables box by clicking on the corresponding selection button. These variables constitute the predictor variables.

Output Variable: Select one variable as the dependent variable from the Variables box by clicking on the corresponding selection button. This is the variable being classified.

Click Next, and the following dialog box appears. Here you specify the architecture for the neural network.

Normalize input data: Normalizing the data (subtracting the mean and dividing by the standard deviation) is important to ensure that the distance measure accords equal weight to each variable -- without normalization, the variable with the largest scale will dominate the measure.

Number of hidden layers: Up to four hidden layers can be specified; see the introduction section for more detail on layers in a neural network (input, hidden and output).

# Nodes: Specify the number of nodes in each hidden layer. Selecting the number of hidden layers and the number of nodes is largely a matter of trial and error.

# Epochs: An epoch is one sweep through all the records in the training set.

Step size for gradient descent: This is the multiplying factor for the error correction during backpropagation; it is roughly equivalent to the learning rate for the neural network. A low value produces slow but steady learning, a high value produces rapid but erratic learning. Values for the step size typically range from 0.1 to 0.9.

Weight change momentum: In each new round of error correction, some memory of the prior correction is retained so that an outlier that crops up does not spoil accumulated learning. The momentum value ranges from 0-2.

Error tolerance: The error in a particular iteration is backpropagated only if it is greater than the error tolerance. Typically error tolerance is a small value in the range 0 to 1. The default value for error tolerance in XLMinerTM is 0.01.

Weight decay: To prevent over-fitting of the network on the training data set a “Weight decay” is used to penalize the error in each iteration. So if e is the error to be back-propagated, instead (e + w*e) is back-propagated where, w is the weight decay, which can be any value in the range 0-1.

Cost Function : XLminer™ provides four options for cost functions -- squared error, cross entropy, Maximum likelihood and perceptron convergence. The user can select the appropriate one.

Hidden layer sigmoid : The output of every hidden node passes through a sigmoid function. Standard sigmoid function is logistic, the range is between 0 and 1. Symmetric sigmoid function is tanh function, the range being -1 to 1.

Output layer sigmoid : Standard sigmoid function is logistic, the range is between 0 and 1. Symmetric sigmoid function is tanh function, the range being -1 to 1.

Click Next, and the following dialog appears.

Score training data: Select this option to show an assessment of the performance of the tree in classifying the training data. The report is displayed according to your specifications - Detailed, Summary and Lift charts.

Score validation data: Select this option to show an assessment of the performance of the tree in classifying the validation data. The report is displayed according to your specifications - Detailed, Summary and Lift charts.

Score Test Data: The options in this group let you apply the model for scoring to the test partition (if one had been created earlier). The option "Score Test Data" is available only if the dataset contains test partition. Select it to apply the model to test data.

Score new Data: The options in this group let you apply the model for scoring to an altogether new data. Specify where the new data is located. See the Example of Discriminant Analysis for detailed instructions on this.

Score New data in database : See the Example of Discriminant Analysis for detailed instructions on this.

Click Finish, and the output will be displayed in a separate sheet.

二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

2005-1-19 00:58:00

不错,的确是EXCEL平台上的一个好分析工具,可惜只能找到DEMO。。。

二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

2011-5-25 11:25:08
謝謝樓主的分享
二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

2013-9-2 05:47:39
Version 12.5 Now Available for Excel 2007 / 2010 / 2013

See  what's new in XLMiner V12.5. In summary, XLMiner now includes:
•Powerful data exploration and visualization features, in additional to its data preparation, data mining, and time series forecasting methods.
•Support for Microsoft's PowerPivot add-in, which handles 'Big Data' and integrates multiple, disparate data sources into one in-memory database inside Excel.
•Support for Excel 2013, and the new PowerPivot add-in that ships with the new Excel
二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

2014-5-26 08:34:36
XLMiner
二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

相关推荐
栏目导航
热门文章
推荐文章

说点什么

分享

扫码加好友,拉您进群
各岗位、行业、专业交流群