Basic Concepts of Logistic Regression

The basic approach is to use the following regression model, employing the notation from Definition 3 of Method of Least Squares for Multiple Regression:

image2124

where the odds function is as given in the following definition.

Odds Function

Definition 1: Odds(E) is the odds that event E occurs, namely

Odds

Where p has a value 0 ≤ p ≤ 1 (i.e. p is a probability value), we can define the odds function as

Odds function

For our purposes, the odds function has the advantage of transforming the probability function, which has values from 0 to 1, into an equivalent function with values between 0 and ∞. When we take the natural log of the odds function, we get a range of values from -∞ to ∞.

Logit Function

Definition 2: The logit function is the log of the odds function, namely logit(E) = ln Odds(E), or

Logit function

Logistic model

Definition 3: Based on the logistic model as described above, we have

logistic regression formula

where π = P(E). It now follows that (see Exponentials and Logs):

image2133

and so

image2134

Here we switch to the model based on the observed sample (and so the π parameter is replaced by its sample estimate p, the βj coefficients are replaced by the sample estimates bj, and the error term ε is dropped). For our purposes, we take E to be the event that the dependent variable y has a value of 1. If y takes only the values 0 or 1, we can think of E as success and the complement E′ of E as failure. This is as for the trials in a binomial distribution.

Just as for the regression model studied in Regression and Multiple Regression, a sample consists of n data elements of the form (yi, xi1, x ,…, xik), but for logistic regression, each yi only takes the value 0 or 1. Now let Ei = the event that yi = 1 and pi = P(Ei). Just as the regression line provides a way to predict the value of the dependent variable y from the values of the independent variables x1, …, xk, for logistic regression we have

image2139

image2140

Note too that since the yi have a proportion distribution, by Property 2 of Proportion Distributionvar(yi) = pi (1 – pi).

Sigmoid curve

In the case where k = 1, we have

image2142

Such a curve has a sigmoid shape, as shown in Figure 1.

Sigmoid curve

Figure 1 – Sigmoid curve for p

The values of b0 and b1 determine the location direction and spread of the curve. The curve is symmetric about the point where x = -b0/b1. In fact, the value of p is 0.5 for this value of x.

When to use logistic regression

Logistic regression is used instead of ordinary multiple regression because the assumptions required for ordinary regression are not met. In particular

  1. The assumption of the linear regression model that the values of y are normally distributed cannot be met since y only takes the values 0 and 1.
  2. The assumption of the linear regression model that the variance of y is constant across values of x (homogeneity of variances) also cannot be met with a binary variable. Since the variance is p(1–p) when 50 percent of the sample consists of 1’s, the variance is .25, its maximum value. As we move to more extreme values, the variance decreases. When p = .10 or .90, the variance is (.1)(.9) = .09, and so as p approaches 1 or 0, the variance approaches 0.
  3. Using the linear regression model, the predicted values will become greater than one or less than zero if you move far enough on the x-axis. Such values are theoretically inadmissible for probabilities.

For the logistics model, the least-squares approach to calculating the values of the coefficients bi cannot be used. Instead, the maximum likelihood techniques, as described below, are employed to find these values.

Odds Ratio

Definition 4: The odds ratio between two data elements in the sample is defined as follows:

Odds ratio

Using the notation px = P(x), the log odds ratio of the estimates is defined as

Log odds ratio

Interpreting the odds ratio

In the case where

image2152

image2153

Thus,
image2154

Furthermore,  for any value of d

image2155

Note too that when x is a dichotomous variable,

image2156

E.g. when x = 0 for male and x = 1 for female, then e^{b_1} represents the odds ratio between males and females. If for example b1 = 2, and we are measuring the probability of getting cancer under certain conditions, then e^{b_1} = 7.4, which would mean that the odds of females getting cancer would be 7.4 times greater than males under the same conditions.

Maximum Log-likelihood

The model we will use is based on the binomial distribution, namely the probability that the sample data occurs as it does is given by

image2160

Taking the natural log of both sides and simplifying we get the following definition.

Definition 5: The log-likelihood statistic is defined as follows:

Log-likelihood statistic

where the yi are the observed values while the pi are the corresponding theoretical values.

Our objective is to find the maximum value of LL assuming that the pi are as in Definition 3. This will enable us to find the values of the bi coordinates. It might be helpful to review Maximum Likelihood Function to better understand the rest of this topic.

Example

Example 1: A sample of 760 people who received doses of radiation between 0 and 1000 rems was made following a recent nuclear accident. Of these 302 died as shown in the table in Figure 2. Actually, each row in the table represents the midpoint of an interval of 100 rems (i.e. 0-100, 100-200, etc.).

Probability odds Excel

Figure 2 – Data for Example 1 plus probability and odds

Let Ei = the event that a person in the ith interval survived. The table also shows the probability P(Ei) and odds Odds(Ei) of survival for a person in each interval. Note that P(Ei) = the percentage of people in interval i who survived and

image2168

In Figure 3 we plot the values of P(Ei) vs. i and ln Odds(Ei) vs. i. We see that the second of these plots is reasonably linear.

Log odds plot

Figure 3 – Plot of probability and ln odds

Given that there is only one independent variable (namely x = # of rems), we can use the following model

Logit one independent variable

Here we use coefficients a and b instead of b0 and b1 just to keep the notation simple.

We show two different methods for finding the values of the coefficients a and b. The first uses Excel’s Solver tool and the second uses Newton’s method. Before proceeding it might be worthwhile to click on Goal Seeking and Solver to review how to use Excel’s Solver tool and Newton’s Method to review how to apply Newton’s Method. We will use both methods to maximize the value of the log-likelihood statistic as defined in Definition 5.

Sample Size

The recommended minimum sample size for logistic regression is given by 10k/q where k = the number of independent variables and q = the smaller of the percentage of cases for y = 0 or y = 1, with a minimum of 100.

For Example 1, k = 1 and q = 302/760 = .397, and so 10k/q = 25.17. Thus a minimum sample of size 100 is recommended.

See Logistic Regression Sample Size for a more refined method of determining the minimum sample size required for logistic regression.

Examples Workbook

Click here to download the Excel workbook with the examples described on this webpage.

References

Howell, D. C. (2010) Statistical methods for psychology (7th ed.). Wadsworth, Cengage Learning.
https://labs.la.utexas.edu/gilden/files/2016/05/Statistics-Text.pdf

Christensen, R. (2013) Logistic regression: predicting counts.
http://stat.unm.edu/~fletcher/SUPER/chap21.pdf

Wikipedia (2012) Logistic regression
https://en.wikipedia.org/wiki/Logistic_regression

Agresti, A. (2013) Categorical data analysis, 3rd Ed. Wiley.
https://mybiostats.files.wordpress.com/2015/03/3rd-ed-alan_agresti_categorical_data_analysis.pdf

85 thoughts on “Basic Concepts of Logistic Regression”

  1. Hi Charles,

    Can the analysis shown above be adapted for a multivariable approach? As in, can we just add in additional independent variables?

    Best,

    Scott

    Reply
  2. Sir,
    A big thank for this great job. It is very useful to understand in deep logistic models.
    I would like to know if raw data in example 1 are available? I would like to show that results with raw data are the same to results with aggregating data.
    Thanks for this job.
    Thierry

    Reply
  3. Dear Charles

    Sorry, I don’t understand understand how we get from the 2nd part of Definition 3, in which P(E)/(1-P(E)) is equal to the exponential of the regression equation to the 3rd part with P(E) on the left part of the equation. I am sure I would have got this once, but I seem to have forgotten the relevant maths. Any help would be gratefully received. Thank you, Mike.

    Reply
    • Mike,
      Let’s start with 1/[1+e^(-z)], which is equal to 1/[1+1/e^z]. Now multiply the numerator and denominator by e^z to get e^z/[e^z+1]. THis proves that e^z/[1+e^z] = 1/[1+e^(-z)], which is the form of the 3rd part of the equation.
      Charles

      Reply
  4. Dr. Charles, good morning. Is it possible to apply the Aikike criterion (AIC), in multiple binary logistic regression models, to see which one has the best fit? and if possible how would you do it?
    Thank you very much

    Dr. Charles, buenos días. Es posible aplicar el criterio de Aikike (AIC), en modelos de regresión logistica binaria multiple, para observar cual tiene mejor ajuste? y si es posible como lo haría?
    Muchas gracias

    Reply
  5. Charles, Thanks again for your great work. In terms of Sample size calculation, when I calculate, actually I don’t know value of q since I have not started the experiment. How can I use formula 10*k/q? Thank you.

    Reply
  6. I am confused about the definition of L. You say it is the probability that “the sample data occurs as it does”.
    But for a given category cat_i, if p_i is the probability to fall in cat_i; and n the number of trials for that category, the probability to observe n_i “success” is
    \binom{n}{n_i}*p_i^n_i*(1-p_i)*(n-n_i) which would give rise to the product of the preceding formulas for i from 1 to k (k being the number of category).
    But that’s not the way you define L. Am I right that here y_i is n_i/n ? Am I missing something? Of what L is the probability, precisely?

    Reply
  7. Charles, Thank you for great add-inn. Working with regression was never so easy.
    Have a one, maybe silly from your experience perspective, question. How I can calculate 95%CI for regression coefficients ? I would like to plot the S curve and it’s lower & upper 95% bounds.
    Thank you in advance for your help.

    Reply
  8. Charles,
    Thanks for the amazing tool! I’ve tried your logit function on one of my data sets and found perfect agreement with other methods of calculating logit models. Very much appreciated!
    I have a question on the applicability for another data set, in which the I have thousands of repeats on non-independent data. For example, we tested n participants on a test with a binary outcome with a number of independent variables (including repeats) in a fully repeated measures design. I’ve converted the binary outcome results into probabilities and done a multiple regression on the probabilities; however, I’m not satisfied with the resultant curve as for certain levels of the independent variables the regression model probability exceeds 1. Is it correct/justifiable to apply a logistic regression approach?

    Reply
      • Sorry that I wasn’t clear. I have data from a laboratory test with human subjects with several independent variables using a fully repeated measures design and a binary dependent variable. I’m unsure if it would be correct to use a logistic regression in this situation. Thanks

        Reply
          • Repeated measures in that all participants are tested under all combinations of independent variables, such that all manipulations are within subjects factors.

          • Andrew,
            Based on what I understand about your situation, it seems like binary logistic regression could be a good choice.
            In a previous comment, you said that the “levels of the independent variables the regression model probability exceeds 1”. This should not happen in logistic regression.
            Charles

  9. I was curious if you could provide some advice. Long forgotten are my ‘out-of-context’ statistics courses and having stumbled about your blog I hope make a comeback!

    I’m looking to analyze the outcome of lots of simulations. At a high level I am comparing success vs failure of the simulation based on several different parameters. I want to get a report on the best parameter set to most likely have successful runs in the future.

    In other words: y = f(x1,x2,x3,…,xn) where y is 0 or 1 and the x values are categorical (i.e., they only take on specific values though not the same number of options).

    It seems logistic regression you described could be what I need but I could really use your 2cents on if I’m barking up the wrong tree. Thank you in advance!

    Reply
  10. Thanks for the extraordinarily useful page. I’ve downloaded the examples-part 2 and am looking at the multinomial logit example 1. I apologize in advance for asking such a basic question, but if I look at the Odds calculated for cured for Dosage 20, Gender 0, we see 0 observed cured.

    It isn’t clear to me how the Odds for cured could be anything other than zero.

    As I say, it is a very basic question, thanks for an excellent site!

    Reply
    • Richard, when you say “if I look at the Odds calculated for cured for Dosage 20, Gender 0, we see 0 observed cured”, are you referring to some worksheet on the website or to some other that you are working on? In general if the number of successes is 0, the the odds will be 0.
      Charles

      Reply
  11. Hi Charles,
    Say we are doing a binary logistic regression of y with respect to each of the separate variables x_1,x_2, x_3, i.e., we are logistically regressing x_1|y , x_2|y and x_3|y ; x_i is the independent variable in each case. Assume all pairs (b_i, b_j) have the same sign combination, i.e., b_i positive, b_j negative, etc.
    Question: Is there a meaningful way of defining a new variable x from x_1,x_2, x_3 , say x=x_1+x_2+x_3 , so that we can regress x|y ? Specifically, here the x_i are measures of efficiency and y is a measure of control, so I want to measure the overall effect of control on efficiency.
    Thanks in Advance.

    Reply
  12. When we deal with a time series data for x(i)s, among which collinearity is highly likely present, what model can I use to draw inference about the probability between x(i) and a binary dependent variable y(i) ?

    Reply
  13. Hi, Charles

    Similar to the discussion on multiple linear regression, is there a discussion addressing outliers and influential observations in undertaking multiple logistic regression? Do you have a recommended suite of Real Statistics features that one should use?

    Regards,
    Rich

    Reply
  14. Dear Charles,
    I have followed your method using my data set, which has a sample size of ~500 and ~150 unique x-values, however only ~30 of the population died. This results in many infinity and negative infinity values for the ln Odds(E) graph. Is it still possible to create this graph and obtain the coefficients for the line of best fit?
    My end goal is to create an ROC curve to find out how good the x-values are at predicting the outcome (lived/died).
    Thanks a lot,
    Ally

    Reply
    • Ally,
      I suggest that you create an example that adheres to the situation that you describe and try logistic regression to see what happens. My guess is that there is a good chance that it will still be possible to create the ROC curve that you are looking for.
      Charles

      Reply
  15. Mr Charles,
    The information you have provided here is truly helpful for someone like me who is studying statistics. Would you mind elaborating little bit how do we get the likelihood statistic as a binomial distribution?Thank you

    Reply
    • Sorry, but I don’t understand your question. Are you looking for the likelihood statistic for the binomial distribution?
      Charles

      Reply
      • Apologies for not being clear with my question.I’m asking about the likelihood function which is used for evaluating the model.The formula which you have used to compare between the predicted probability and the actual outcome and which is converted to log likelihood statistic which is being maximized.The formula mentioned above definition 5 on this page.

        Reply
        • Chirag,

          I suggest that you look at the following webpage
          https://real-statistics.com/general-properties-of-distributions/maximum-likelihood-function/

          The idea is that the probability that the sample will occur as it does is a product of probabilities. This is because the sample elements are independent of each other; i.e. P(A and B) = P(A) x P(B) when A and B are independent events. Thus the likelihood function L is a product of probability function values (that are dependent on certain parameters). For logistic regression, the probability function is the pdf for the binary distribution.

          Since the sample that was observed actually did occur, the approach we use is to find the values of the parameters that maximize L(i.e. that make the sample events most likely)

          Charles

          Reply
  16. Dear Charles,
    Greetings. As a novice, I must say I’ve found your website immensely helpful.
    I have the following data which I need to build a regression model for. Kindly suggest how I might go about it…

    Dependent Variable: Binary Categorical
    Independent Variables: Many are categorical and many are interval

    I need to see if it is possible to build a model and test it. Kindly advise.

    Reply
  17. I am so sorry for the last respond, it might be confusing because I forgot to put ‘0’
    below is the data I am working on:

    age leave stay total
    18.6 0 2 2
    18.8 2 0 2
    18.9 1 0 1
    19 1 0 1
    19.1 0 1 1
    19.2 2 1 3
    19.3 0 1 1
    19.4 0 1 1
    19.5 1 0 1
    19.6 0 2 2
    19.8 4 2 6
    19.9 1 1 2
    20 2 2 4
    20.1 3 4 7
    20.2 1 3 4
    20.3 4 1 5
    20.4 5 9 14
    20.5 3 4 7
    20.6 3 7 10
    20.7 4 6 10
    20.8 4 8 12
    20.9 5 16 21
    21 7 8 15
    21.1 5 10 15
    21.2 5 13 18
    21.3 7 16 23
    21.4 6 17 23
    21.5 14 13 27
    21.6 12 13 25
    21.7 12 17 29
    21.8 17 9 26
    21.9 7 23 30
    22 20 32 52
    22.1 19 32 51
    22.2 21 49 70
    22.3 23 55 78
    22.4 25 36 61
    22.5 27 46 73
    22.6 22 54 76
    22.7 32 31 63
    22.8 32 50 82
    22.9 50 39 89
    23 39 65 104
    23.1 40 81 121
    23.2 42 89 131
    23.3 44 80 124
    23.4 71 66 137
    23.5 50 84 134
    23.6 64 94 158
    23.7 57 95 152
    23.8 59 108 167
    23.9 77 88 165
    24 61 105 166
    24.1 66 118 184
    24.2 70 131 201
    24.3 73 114 187
    24.4 66 108 174
    24.5 64 137 201
    24.6 62 159 221
    24.7 66 135 201
    24.8 86 130 216
    24.9 96 126 222
    25 82 152 234
    25.1 75 135 210
    25.2 88 143 231
    25.3 72 149 221
    25.4 92 156 248
    25.5 89 171 260
    25.6 95 173 268
    25.7 80 172 252
    25.8 103 150 253
    25.9 94 183 277
    26 92 156 248
    26.1 95 185 280
    26.2 97 196 293
    26.3 86 179 265
    26.4 95 208 303
    26.5 72 187 259
    26.6 94 169 263
    26.7 89 195 284
    26.8 84 167 251
    26.9 86 157 243
    27 83 181 264
    27.1 90 177 267
    27.2 75 187 262
    27.3 76 191 267
    27.4 75 206 281
    27.5 72 197 269
    27.6 66 207 273
    27.7 84 150 234
    27.8 77 174 251
    27.9 59 152 211
    28 67 166 233
    28.1 64 164 228
    28.2 64 174 238
    28.3 65 165 230
    28.4 82 200 282
    28.5 80 195 275
    28.6 78 179 257
    28.7 83 155 238
    28.8 89 148 237
    28.9 74 150 224
    29 68 153 221
    29.1 69 125 194
    29.2 71 180 251
    29.3 67 156 223
    29.4 55 187 242
    29.5 75 164 239
    29.6 63 151 214
    29.7 58 164 222
    29.8 58 143 201
    29.9 64 146 210
    30 62 111 173
    30.1 55 159 214
    30.2 66 160 226
    30.3 54 147 201
    30.4 49 149 198
    30.5 58 141 199
    30.6 50 163 213
    30.7 50 179 229
    30.8 51 143 194
    30.9 52 129 181
    31 56 136 192
    31.1 46 138 184
    31.2 40 134 174
    31.3 65 147 212
    31.4 49 145 194
    31.5 58 125 183
    31.6 34 184 218
    31.7 51 136 187
    31.8 43 125 168
    31.9 47 116 163
    32 43 129 172
    32.1 43 115 158
    32.2 52 142 194
    32.3 41 114 155
    32.4 42 132 174
    32.5 61 137 198
    32.6 45 157 202
    32.7 50 127 177
    32.8 43 117 160
    32.9 50 114 164
    33 49 120 169
    33.1 32 129 161
    33.2 35 117 152
    33.3 51 134 185
    33.4 41 131 172
    33.5 49 131 180
    33.6 40 154 194
    33.7 39 101 140
    33.8 36 135 171
    33.9 45 108 153
    34 42 92 134
    34.1 41 115 156
    34.2 48 114 162
    34.3 40 117 157
    34.4 38 132 170
    34.5 40 114 154
    34.6 25 134 159
    34.7 27 110 137
    34.8 52 96 148
    34.9 35 106 141
    35 38 89 127
    35.1 23 103 126
    35.2 38 90 128
    35.3 34 101 135
    35.4 34 106 140
    35.5 38 108 146
    35.6 45 112 157
    35.7 25 103 128
    35.8 37 91 128
    35.9 22 84 106
    36 31 93 124
    36.1 37 81 118
    36.2 18 90 108
    36.3 39 91 130
    36.4 26 96 122
    36.5 30 98 128
    36.6 33 102 135
    36.7 28 71 99
    36.8 27 74 101
    36.9 28 64 92
    37 33 74 107
    37.1 19 78 97
    37.2 25 81 106
    37.3 31 72 103
    37.4 27 88 115
    37.5 31 89 120
    37.6 25 83 108
    37.7 25 69 94
    37.8 16 88 104
    37.9 28 64 92
    38 19 43 62
    38.1 23 54 77
    38.2 25 73 98
    38.3 21 71 92
    38.4 19 76 95
    38.5 21 64 85
    38.6 17 70 87
    38.7 27 68 95
    38.8 21 63 84
    38.9 18 66 84
    39 16 54 70
    39.1 22 59 81
    39.2 20 72 92
    39.3 14 45 59
    39.4 17 69 86
    39.5 19 72 91
    39.6 16 65 81
    39.7 12 64 76
    39.8 16 51 67
    39.9 12 42 54
    40 15 59 74
    40.1 19 41 60
    40.2 10 59 69
    40.3 10 57 67
    40.4 18 60 78
    40.5 17 77 94
    40.6 17 51 68
    40.7 13 60 73
    40.8 13 51 64
    40.9 19 52 71
    41 10 66 76
    41.1 22 61 83
    41.2 8 63 71
    41.3 19 37 56
    41.4 16 51 67
    41.5 13 56 69
    41.6 15 75 90
    41.7 9 54 63
    41.8 19 56 75
    41.9 14 50 64
    42 16 51 67
    42.1 15 52 67
    42.2 13 59 72
    42.3 12 56 68
    42.4 22 62 84
    42.5 16 67 83
    42.6 13 84 97
    42.7 9 82 91
    42.8 15 67 82
    42.9 19 63 82
    43 19 70 89
    43.1 21 74 95
    43.2 17 83 100
    43.3 12 80 92
    43.4 17 73 90
    43.5 13 73 86
    43.6 10 75 85
    43.7 15 72 87
    43.8 13 69 82
    43.9 9 59 68
    44 12 69 81
    44.1 12 79 91
    44.2 14 71 85
    44.3 12 87 99
    44.4 9 72 81
    44.5 11 82 93
    44.6 14 83 97
    44.7 14 68 82
    44.8 14 55 69
    44.9 12 77 89
    45 11 72 83
    45.1 7 63 70
    45.2 13 61 74
    45.3 9 65 74
    45.4 12 64 76
    45.5 12 75 87
    45.6 15 71 86
    45.7 7 68 75
    45.8 9 47 56
    45.9 13 57 70
    46 10 41 51
    46.1 12 46 58
    46.2 12 43 55
    46.3 16 44 60
    46.4 9 55 64
    46.5 6 45 51
    46.6 12 41 53
    46.7 10 47 57
    46.8 3 36 39
    46.9 8 44 52
    47 8 34 42
    47.1 9 35 44
    47.2 13 41 54
    47.3 4 51 55
    47.4 11 41 52
    47.5 7 31 38
    47.6 3 55 58
    47.7 3 43 46
    47.8 8 37 45
    47.9 8 30 38
    48 8 36 44
    48.1 7 33 40
    48.2 6 31 37
    48.3 5 27 32
    48.4 6 35 41
    48.5 6 38 44
    48.6 9 38 47
    48.7 4 27 31
    48.8 5 37 42
    48.9 9 29 38
    49 6 25 31
    49.1 10 24 34
    49.2 8 35 43
    49.3 6 31 37
    49.4 5 33 38
    49.5 5 27 32
    49.6 6 31 37
    49.7 4 30 34
    49.8 6 30 36
    49.9 7 27 34
    50 2 20 22
    50.1 13 24 37
    50.2 4 32 36
    50.3 3 28 31
    50.4 3 22 25
    50.5 1 32 33
    50.6 1 24 25
    50.7 5 24 29
    50.8 4 20 24
    50.9 4 22 26
    51 1 14 15
    51.1 2 23 25
    51.2 3 21 24
    51.3 4 16 20
    51.4 3 16 19
    51.5 3 22 25
    51.6 5 20 25
    51.7 3 14 17
    51.8 2 20 22
    51.9 4 11 15
    52 3 18 21
    52.1 4 12 16
    52.2 2 11 13
    52.3 2 13 15
    52.4 2 14 16
    52.5 7 11 18
    52.6 4 13 17
    52.7 1 25 26
    52.8 5 6 11
    52.9 2 11 13
    53 5 11 16
    53.1 1 7 8
    53.2 2 11 13
    53.3 1 14 15
    53.4 2 28 30
    53.5 6 9 15
    53.6 3 10 13
    53.7 2 13 15
    53.8 5 8 13
    53.9 1 8 9
    54 0 8 8
    54.1 1 12 13
    54.2 0 3 3
    54.3 0 12 12
    54.4 0 7 7
    54.5 2 6 8
    54.6 1 7 8
    54.7 0 12 12
    54.8 0 9 9
    54.9 0 5 5
    55 8 3 11
    55.1 49 1 50
    55.6 1 1 2
    55.7 0 1 1
    55.8 1 1 2
    55.9 1 0 1
    56.1 1 0 1
    56.4 0 1 1
    56.5 0 2 2
    56.6 0 1 1
    56.7 1 0 1
    56.9 0 1 1
    57 2 0 2
    57.1 1 0 1
    57.3 0 1 1
    57.8 0 1 1
    58.5 0 1 1
    58.7 1 0 1
    60.6 1 0 1
    61.2 1 0 1
    70.7 0 1 1
    Grand Total 10476 27764 38240

    Thank you for your kindness, Sir.
    Regard,

    Reply
    • Please send me an Excel file with your data by email along with the results that you obtained that you say are different from mine. Please make sure that when you use the Real Statistics software that you don’t include the totals for each item in the input data range.
      Charles

      Reply
  18. I wish to used this model to predict the attrition probability, but I wonder when seeing my P(E) and Odds(E) plot because it seems different from yours, do I need to check the data plot to decide whether my data could be modeling as binary logistic regression? if so, how to check it?

    Reply
    • When you say that your P(E) and Odds(E) are different from mine, do you mean that you got a different answer for one of the examples on the website?
      Charles

      Reply
      • I am not using the example data, but I use mine, in the picture your P(E) and Odds(E) probability linearly decreasing by the x- axis getting bigger, but my data doesn’t, so my point is, is there any requirement when choosing the binary logistic regression as a model?

        Reply
        • The assumptions are: the dependent variable has exactly two outcomes; observations are independent; no collinearity, outliers or high leverage/influencer values; the model fits (i.e. there is a linear relationship between the independent variables and the logit transformation of the dependent variable).

          If you send me an Excel file with your data, I will try to figure out what is happening.

          Charles

          Reply
    • The referenced webpage describes how to get the P(E) values. The webpage sometimes uses the notation p for P(E). The other webpages on the Logistic Regression topic gets into more detail. You can also download the spreadsheets with all the examples shown on the website to get even more information.
      Charles

      Reply
  19. I tried the logistic regression tool with a data-set of Rows-142829 and Columns- 47 and my system hangs each time. Please can you let me know what is the capability of the addin. Also recommend me some way i can apply logistic regression on the same.

    Reply
    • Assuming that the data is in raw format (and not summary format), there are two possible approaches:

      1. Try using the Logistic Regression data analysis tool, making sure that Raw data option is selected, but the Show summary in output option is not selected. To use the summary output option the number of rows must be less than 65,000.

      2. If the above approach doesn’t work, use the LogitMatches function as described on the webpage https://real-statistics.com/logistic-regression/real-statistics-functions-logistic-regression/

      With this amount of data, the analysis will be quite slow.

      Charles

      Reply
  20. Dear Charles,

    I was wondering how the sigmoid curve is plotted. It seems that the probabilty p given by p=1/(1+EXP(-B0-B1*x)) is plotted on the y-axis, what did you use on the x-axis? Seems to be from -3 to 3

    Thanks in advance

    Reply
  21. Hi Charles, I’m trying to predict how suitable a prospective client would be for me; based on historical of my already existing clients. This will be based on 5 independent categorical variables.What do you suggest a suitable regression techniques would be?

    Reply
    • You have several choices. First you need to decide whether your dependent variable is categorical or not.

      If it is categorical then some version of logistic regression could be used. If there are two outcomes then the model is a binary logistic regression model (e.g. client is suitable or client is unsuitable). If there are multiple categories then you could use multinomial logistic regression or more likely ordinal logistic regression (e.g. client is a top prospect, client is a good prospect, client is a fair prospect or client is not a suitable prospect).

      With only categorical variables you could also use log-linear models.

      With a quantitative dependent variable (e.g. client is assigned a numeric rating from 0 to 100, with 100 representing clients that are most suitable), you could multiple regression employing dummy variables to handle the categorical independent variables.

      All of these approaches are described on the Real Statistics website.

      Charles

      Reply
  22. Dear Sir,
    Thank You for your great explanations. Logic regression is used for item response theory also. Do you have any explanations about that?

    Reply
    • Venura,
      I don’t have any further information about this subject on the website. The site is evolving all the time and new information is constantly being added. What sort of explanations on this subject would be most helpful to you?
      Charles

      Reply
        • Sorry Venura, I should have read your question more carefully. I now see that you are referencing item response theory. This is a topic that I need to add to the website.
          Charles

          Reply
  23. In example 1, there’s a plot of the Odds(E). The plot shown looks to me like it is not the Odds(E), but rather the logit = ln Odds(E). Is this correct?

    Reply
  24. Sir

    I am at little confused about the value of yi in definition 5. In definition 5 you said yi is observed value. But I think if it is a binomial distribution y can only take the value of either 1 or 0, how could it be observed value of p?

    Colin

    Reply
    • Colin,
      yi only takes a value of 0 or 1. pi is the predicted value, which can take any value between 0 and 1. Of course we will use the pi as a way of estimating the yi (even for unobserved yi). If pi > .5 (or some other preassigned value) then pi predicts a value for yi of 1 and pi < .5 predicts a value for yi of 0. Charles

      Reply
      • Sir

        I am sorry. I still don’t understand. In the “Finding Logistic Regression Coefficients using Excel’s Solver” you said yi “is the observed probability of survival in the ith of r intervals” and the value of yi in Figure 1 of “Finding Logistic Regression Coefficients using Excel’s Solver” does not take the value of either 0 or 1, which makes me confused.

        Colin

        Reply
        • Colin,
          You are correct. It is a little confusing. I should have said that yi = the fraction of subjects in the ith interval that survived. I have now made this revision on the webpage.
          Charles

          Reply

Leave a Reply to ikhlas sadimin Cancel reply