When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. Logistic regression - Wikipedia

    en.wikipedia.org/wiki/Logistic_regression

    x M and, as in the example above, two categorical values (y = 0 and 1). For the simple binary logistic regression model, we assumed a linear relationship between the predictor variable and the log-odds (also called logit) of the event that =. This linear relationship may be extended to the case of M explanatory variables:

  3. Hosmer–Lemeshow test - Wikipedia

    en.wikipedia.org/wiki/Hosmer–Lemeshow_test

    6. Calculate the p-value Compare the computed Hosmer–Lemeshow statistic to a chi-squared distribution with Q − 2 degrees of freedom to calculate the p-value. There are Q = 10 groups in the caffeine example, giving 10 – 2 = 8 degrees of freedom. The p-value for a chi-squared statistic of 17.103 with df = 8 is p = 0.029. The p-value is ...

  4. Pseudo-R-squared - Wikipedia

    en.wikipedia.org/wiki/Pseudo-R-squared

    R 2 L is given by Cohen: [1] =. This is the most analogous index to the squared multiple correlations in linear regression. [3] It represents the proportional reduction in the deviance wherein the deviance is treated as a measure of variation analogous but not identical to the variance in linear regression analysis. [3]

  5. Linear predictor function - Wikipedia

    en.wikipedia.org/wiki/Linear_predictor_function

    The basic form of a linear predictor function () for data point i (consisting of p explanatory variables), for i = 1, ..., n, is = + + +,where , for k = 1, ..., p, is the value of the k-th explanatory variable for data point i, and , …, are the coefficients (regression coefficients, weights, etc.) indicating the relative effect of a particular explanatory variable on the outcome.

  6. Logistic distribution - Wikipedia

    en.wikipedia.org/wiki/Logistic_distribution

    In probability theory and statistics, the logistic distribution is a continuous probability distribution. Its cumulative distribution function is the logistic function, which appears in logistic regression and feedforward neural networks. It resembles the normal distribution in shape but has heavier tails (higher kurtosis).

  7. One in ten rule - Wikipedia

    en.wikipedia.org/wiki/One_in_ten_rule

    In statistics, the one in ten rule is a rule of thumb for how many predictor parameters can be estimated from data when doing regression analysis (in particular proportional hazards models in survival analysis and logistic regression) while keeping the risk of overfitting and finding spurious correlations low. The rule states that one ...

  8. Autoregressive moving-average model - Wikipedia

    en.wikipedia.org/wiki/Autoregressive_moving...

    The notation ARMAX(p, q, b) refers to a model with p autoregressive terms, q moving average terms and b exogenous inputs terms. The last term is a linear combination of the last b terms of a known and external time series d t {\displaystyle d_{t}} .

  9. Conditional logistic regression - Wikipedia

    en.wikipedia.org/.../Conditional_logistic_regression

    Conditional logistic regression is an extension of logistic regression that allows one to account for stratification and matching. Its main field of application is observational studies and in particular epidemiology. It was devised in 1978 by Norman Breslow, Nicholas Day, Katherine Halvorsen, Ross L. Prentice and C. Sabai. [1]