常柴股份有限公司章程_第1页
常柴股份有限公司章程_第2页
常柴股份有限公司章程_第3页
常柴股份有限公司章程_第4页
常柴股份有限公司章程_第5页
已阅读5页,还剩37页未读 继续免费阅读

下载本文档

版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领

文档简介

Slide 1,The Kleinbaum Sample Problem,This problem comes from an example in the text: David G. Kleinbaum. Logistic Regression: A Self-Learning Text. New York: Springer-Verlag, 1966. Pages 256-257.The problem is to examine the relationship between the dependent variable Confidence in the legal system, (CONFIDEN) and three independent variables: Social class (CLASS), Number of times victimized (VICTIM), and age (AGE).Confidence in the legal system, the dependent variable, is metric so we could use multiple regression analysis. However, the author opts to convert Confidence in the legal system to a dichotomous variable by dividing the score above and below the median value of 10. A new dependent variable, High confidence in the legal system (HIGHCONF) was created where 1 stands for high confidence and 0 stands for low confidence.The data for this problem is: ConfidenceInLegalSystem.Sav.,Kleinbaum Logistic Regression Problem,Slide 2,Stage One: Define the Research Problem,In this stage, the following issues are addressed:Relationship to be analyzedSpecifying the dependent and independent variablesMethod for including independent variables,Kleinbaum Logistic Regression Problem,Relationship to be analyzed,The problem is to examine the relationship between the dependent variable Confidence in the legal system, (CONFIDEN) and three independent variables: Social class (CLASS), Number of times victimized (VICTIM), and age (AGE).,Slide 3,Specifying the dependent and independent variables,The dependent variable, High confidence in the legal system (HIGHCONF) was created so that 1 stands for high confidence and 0 stands for low confidence.The independent variables are:CLASS Social class statusVICTIM Number of times victimizedAGE Age of respondentCLASS Social class status is a nonmetric variable with three response options: 1 = Low, 2 = Medium, and 3 = High. While a case could be made that it can be treated as a scale variable, we will treat it as nonmetric and use the SPSS facility in logistic regression to enter it as a categorical variable.VICTIM Number of times victimized has a range from 0 to 2.,Kleinbaum Logistic Regression Problem,Method for including independent variables,Since we are interested in the relationship between the dependent variable and all of the independent variables, we will use direct entry of the independent variables.,Slide 4,Stage 2: Develop the Analysis Plan: Sample Size Issues,In this stage, the following issues are addressed:Missing data analysisMinimum sample size requirement: 15-20 cases per independent variable,Kleinbaum Logistic Regression Problem,Missing data analysis,There is no missing data in this problem.,Minimum sample size requirement: 15-20 cases per independent variable,The CLASS Social class status variable has three categories, so dummy coding it will require two variables, bringing the total number of independent variables is 4. The data set has 39 cases and 4 independent variables for a ratio of 10 to 1, falling short of the requirement that we have 15-20 cases per independent variable.,Slide 5,Stage 2: Develop the Analysis Plan: Measurement Issues:,In this stage, the following issues are addressed:Incorporating nonmetric data with dummy variablesRepresenting Curvilinear Effects with PolynomialsRepresenting Interaction or Moderator Effects,Kleinbaum Logistic Regression Problem,Incorporating Nonmetric Data with Dummy Variables,The logistic regression procedure will dummy code the nonmetric variables for us.,Representing Curvilinear Effects with Polynomials,We do not have any evidence of curvilinear effects at this point in the analysis.,Representing Interaction or Moderator Effects,We do not have any evidence at this point in the analysis that we should add interaction or moderator variables.,Slide 6,Stage 3: Evaluate Underlying Assumptions,In this stage, the following issues are addressed:Nonmetric dependent variable with two groupsMetric or dummy-coded independent variables,Kleinbaum Logistic Regression Problem,Nonmetric dependent variable having two groups,The dependent variable HIGHCONF High confidence in legal system is a dichotomous variable.,Metric or dummy-coded independent variables,The independent variable CLASS Social class status is nonmetric and will be recoded into two dichotomous variables automatically using an SPSS option for designating an independent variable as categorical.The independent variables VICTIM Number of times victimized and AGE Age of respondent are metric variables.,Slide 7,Stage 4: Estimation of Logistic Regression and Assessing Overall Fit: Model Estimation,In this stage, the following issues are addressed:Compute logistic regression model,Kleinbaum Logistic Regression Problem,Compute the logistic regression,The steps to obtain a logistic regression analysis are detailed on the following screens.,Slide 8,Requesting a Logistic Regression,Kleinbaum Logistic Regression Problem,Slide 9,Specifying the Dependent Variable,Kleinbaum Logistic Regression Problem,Slide 10,Specifying the Independent Variables,Kleinbaum Logistic Regression Problem,Slide 11,Specify the Categorical Independent Variable,Kleinbaum Logistic Regression Problem,Slide 12,Specify the method for entering variables,Kleinbaum Logistic Regression Problem,Slide 13,Specifying Options to Include in the Output,Kleinbaum Logistic Regression Problem,Slide 14,Specifying the New Variables to Save,Kleinbaum Logistic Regression Problem,Slide 15,Complete the Logistic Regression Request,Kleinbaum Logistic Regression Problem,Slide 16,Stage 4: Estimation of Logistic Regression and Assessing Overall Fit: Assessing Model Fit,In this stage, the following issues are addressed:Significance test of the model log likelihood (Change in -2LL)Measures Analogous to R: Cox and Snell R and Nagelkerke RHosmer-Lemeshow Goodness-of-fitClassification matricesCheck for Numerical ProblemsPresence of outliers,Kleinbaum Logistic Regression Problem,Categorical variable recoding,At the start of the output, SPSS reports how it dummy coded the variable CLASS Social class status:,SPSS does not assign new names to the dummy coded variable, instead it will refer to the variables as CLASS(1) and CLASS(2). CLASS(1) corresponds to Lower Class; CLASS(2) corresponds to Middle Class, and Upper Class is the omitted category.,Slide 17,Initial statistics before independent variables are included,The Initial Log Likelihood Function, (-2 Log Likelihood or -2LL) is a statistical measure like total sums of squares in regression. If our independent variables have a relationship to the dependent variable, we will improve our ability to predict the dependent variable accurately, and the log likelihood value will decrease. The initial 2LL value is 54.040 on step 0, before any variables have been added to the model.,Kleinbaum Logistic Regression Problem,Slide 18,Significance test of the model log likelihood,The difference between these two measures is the model child-square value (17.863 = 54.040 36.177) that is tested for statistical significance. This test is analogous to the F-test for R or change in R value in multiple regression which tests whether or not the improvement in the model associated with the additional variables is statistically significant.,In this problem the model Chi-Square value of 17.863 has a significance of 0.001, less than 0.05, so we conclude that there is a significant relationship between the dependent variable and the set of independent variables.,Kleinbaum Logistic Regression Problem,Slide 19,Measures Analogous to R,The next SPSS outputs indicate the strength of the relationship between the dependent variable and the independent variables, analogous to the R measures in multiple regression.,The relationship between the dependent variable and independent variables is strong, indicated by the value of Nagelkerke R2 which is 0.490. Using the interpretive criteria for R, we would characterize this relationship as strong.,Kleinbaum Logistic Regression Problem,Slide 20,Correspondence of Actual and Predicted Values of the Dependent Variable,The final measure of model fit is the Hosmer and Lemeshow goodness-of-fit statistic, which measures the correspondence between the actual and predicted values of the dependent variable. In this case, better model fit is indicated by a smaller difference in the observed and predicted classification. A good model fit is indicated by a nonsignificant chi-square value.,The goodness-of-fit measure has a value of 5.507 which has the desirable outcome of nonsignificance.,Kleinbaum Logistic Regression Problem,Slide 21,The Classification Matrices,The classification matrices in logistic regression serve the same function as the classification matrices in discriminant analysis, i.e. evaluating the accuracy of the model.,If the predicted and actual group memberships are the same, i.e. 1 and 1 or 0 and 0, then the prediction is accurate for that case. If predicted group membership and actual group membership are different, the model misses for that case. The overall percentage of accurate predictions (71.8% in this case) is the measure of a model that I rely on most heavily for this analysis as well as for discriminant analysis because it has a meaning that is readily communicated, i.e. the percentage of cases for which our model predicts accurately. To evaluate the accuracy of the model, we compute the proportional by chance accuracy rate and the maximum by chance accuracy rates, if appropriate. The proportional by chance accuracy rate is equal to 0.500 (0.4872 + 0.5132). A 25% increase over the proportional by chance accuracy rate would equal 0.625. Our model accuracy race of 71.79% exceeds this criterion.With 51% of the cases in one group and 49% of the cases in the other group, we do not have a dominant category that would require us to compare our results to the maximum by chance accuracy rate.,Slide 22,The Stacked Histogram,SPSS provides a visual image of the classification accuracy in the stacked histogram as shown below. To the extent to which the cases in one group cluster on the left and the other group clusters on the right, the predictive accuracy of the model will be higher.As we can see in this plot, there is some overlapping between the two groups.,Slide 23,Check for Numerical Problems,There are several numerical problems that can in logistic regression that are not detected by SPSS or other statistical packages: multicollinearity among the independent variables, zero cells for a dummy-coded independent variable because all of the subjects have the same value for the variable, and complete separation whereby the two groups in the dependent event variable can be perfectly separated by scores on one of the independent variables.All of these problems produce large standard errors (over 2) for the variables included in the analysis and very often produce very large B coefficients as well. If we encounter large standard errors for the predictor variables, we should examine frequency tables, one-way ANOVAs, and correlations for the variables involved to try to identify the source of the problem.,The standard errors and B coefficients are not excessively large, so there is no evidence of a numeric problem with this analysis.,Kleinbaum Logistic Regression Problem,Slide 24,There are two outputs to alert us to outliers that we might consider excluding from the analysis: listing of residuals and saving Cooks distance scores to the data set.SPSS provides a casewise list of residuals that identify cases whose residual is above or below a certain number of standard deviation units. Like multiple regression there are a variety of ways to compute the residual. In logistic regression, the residual is the difference between the observed probability of the dependent variable event and the predicted probability based on the model. The standardized residual is the residual divided by an estimate of its standard deviation. The deviance is calculated by taking the square root of -2 x the log of the predicted probability for the observed group and attaching a negative sign if the event did not occur for that case. Large values for deviance indicate that the model does not fit the case well. The studentized residual for a case is the change in the model deviance if the case is excluded. Discrepancies between the deviance and the studentized residual may identify unusual cases. (See the SPSS chapter on Logistic Regression Analysis for additional details).In the output for our problem, SPSS informs us that there is one outlier in this analysis:,Presence of outliers,Kleinbaum Logistic Regression Problem,Slide 25,Cooks Distance,SPSS has an option to compute Cooks distance as a measure of influential cases and add the score to the data editor. I am not aware of a precise formula for determining what cutoff value should be used, so we will rely on the more traditional method for interpreting Cooks distance which is to identify cases that either have a score of 1.0 or higher, or cases which have a Cooks distance substantially different from the other. The prescribed method for detecting unusually large Cooks distance scores is to create a scatterplot of Cooks distance scores versus case id.,Kleinbaum Logistic Regression Problem,Slide 26,Request the Scatterplot,Kleinbaum Logistic Regression Problem,Slide 27,Specifying the Variables for the Scatterplot,Kleinbaum Logistic Regression Problem,Slide 28,The Scatterplot of Cooks Distances,On the plot of Cooks distances shown below, we see no cases that exceed the 1.0 rule of thumb for influential cases. We do, however, identify cases that have relatively larger Cooks distance values (above 0.6) than the majority of cases. However, with the small sample size we have in this problem, I am not inclined to remove any cases unless they were extreme outliers or influential cases.,Kleinbaum Logistic Regression Problem,Slide 29,Stage 5: Interpret the Results,In this section, we address the following issues:Identifying the statistically significant predictor variablesDirection of relationship and contribution to dependent variable,Kleinbaum Logistic Regression Problem,Slide 30,Identifying the statistically significant predictor variables,The coefficients are found in the column labeled B, and the test that the coefficient is not zero, i.e. changes the odds of the dependent variable event is tested with the Wald statistic, instead of the t-test as was done for the individual B coefficients in the multiple regression equation.,As shown above, only the variables VICTIM Number of times victimized and AGE Age of respondent have a statistically significant individual relationship with the dependent variable.,Kleinbaum Logistic Regression Problem,Slide 31,The predictor variable with the strongest relationship is VICTIM. The negative sign of the B coefficient and the value of Exp(B) less than 1.0 both indicate that the relationship is inverse: the more times one is victimized, the less likely one is to have confidence in the legal system.With the inverse relationship, it may make more sense to invert the odds ratio (1 / odds ratio) and interpret the probability of not belonging to the dependent variable group assigned the code of 1. In this problem, we could say that every time a person is victimized, they are 4.2 times less likely to have high confidence in the legal system (1/.236 = 4.2).Age has a direct relationship with confidence in the legal system: as one gets older, ones confidence in the legal system increases. For every 1 year increase in age, the odds of having high confidence in the legal system increase 1.2 times.,Direction of relationship and contribution to dependent variable,Kleinbaum Logistic Regression Problem,Slide 32,Stage 6: Validate The Model,When we have a small sample in the full data set as we do in this problem, a split half validation analysis is almost guaranteed to fail because we will have little power to detect statistical differences in analyses of the validation samples. In this circumstance, our alternative is to conduct validation analyses with random samples that comprise the majority of the sample. We will demonstrate this procedure in the following steps:Computing the First Validation AnalysisComputing the Second Validation AnalysisThe Output for the Validation Analysis,Kleinbaum Logistic Regression Problem,Computing the First Validation Analysis,We set the random number seed and modify ou

温馨提示

  • 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
  • 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
  • 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
  • 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
  • 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
  • 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
  • 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

评论

0/150

提交评论