Dsom 6 adaptive prac.
True or false: R2 can decrease as we add more predictor variables to the linear regression model True false question. True False
False
In the presence of changing variability, the OLS estimators are 'BLANK', but their estimated standard errors are inappropriate
unbiased
A dummy variable, also referred to as an indicator or a binary variable, takes on numerical values of 1 or 0 to describe two categories of a categorical variable.For a predictor variable that is a dummy variable, it is common to refer to the category that assumes a value of 0 as: Please select all that apply. Multiple select question. Dummy category Regression dummy Benchmark category Reference category
Benchmark category Reference category
The nonzero slope coefficient test is used for a renowned financial application referred to as the capital asset pricing model (CAPM).The model y = α + βx + ɛ, is essentially a simple linear regression model that uses α and β, in place of the usual β0 and β1, to represent the intercept and the slope coefficients, respectively. Which of the following is true about the slope coefficient α, called the stock's alpha? Select that apply! Multiple select question. Measures how sensitive the stock's return is to changes in the level of the overall market Abnormal returns are positive when α > 0 Abnormal returns are negative when α < 0. The CAPM theory predicts α to be zero
Abnormal returns are positive when α > 0 Abnormal returns are negative when α < 0. The CAPM theory predicts α to be zero
In the presence of correlated observations, the OLS estimators are unbiased, but their estimated standard errors are inappropriate.Which of the following could happen as a result? Choose all that apply! Multiple choice question. All of the answers are correct The t test may suggest that the predictor variables are individually and jointly significant when this is not true The model looks better than it really is with a spuriously high R2 The F test may suggest that the predictor variables are individually and jointly significant when this is not true
All of the answers are correct
When confronted with multicollinearity, the best approach may be to do 'BLANK' if the estimated model yields a high R2,
nothing
The assumption of constant variability of observations often breaks down in studies with cross-sectional data. Consider the model y = β0 + β1x + ɛ, where y is a household's consumption expenditure and x is its disposable income. It may be unreasonable to assume that the variability of consumption is the same across a cross-section of household incomes. This violation is called: Multiple choice question. Multicollinearity Correlated Observations Changing variability Nonlinear Patterns
Changing variability
Which of the following are the assumptions that underlie the classical linear regression model? Please select all that apply! Multiple select question. Conditional on x1, x2,.., xk, the error term ɛ is uncorrelated across observations; or, in statistical terminology, there is no serial correlation. There is an exact linear relationship among the predictor variables; or, in statistical terminology, there is no perfect multicollinearity. The regression model given by y = β0 + β1x1 + β2x2 +... + βkxk + ɛ is linear in the parameters β0, β1,..., βk. The error term ɛ is correlated with any of the predictor variables x1, x2,..., xk
Conditional on x1, x2,.., xk, the error term ɛ is uncorrelated across observations; or, in statistical terminology, there is no serial correlation. The regression model given by y = β0 + β1x1 + β2x2 +... + βkxk + ɛ is linear in the parameters β0, β1,..., βk.
We can use residual plots to gauge changing variability. The residuals are generally plotted against each predictor variable xj. There is a violation if the variability increases or 'BLANK' over the values of xj.
Decreases
The example of momentum p is the product of the mass m and the velocity v of an object; that is, p = mv, is an example of a 'BLANK' relationship.
Deterministic
If the value of the response variable is uniquely determined by the values of the predictor variables, we say that the relationship between the variables is: (Choose the correct response) Multiple choice question. Random Stochastic Deterministic Regressive
Deterministic; The relationship between the response variable and the predictor variables is deterministic
For the linear regression model, y = β0 + β1x1 + β2x2 + . . . + βkxk + ɛ, which of the following are the competing hypotheses used for a test of joint significance? Choose both the correct test for the null and alternative hypotheses. Multiple select question. HA: βj ≠βj0 H0:β1=β2=... =βk=0 H0: βj =βj0 HA:At least one βi≠0
H0:β1=β2=... =βk=0 HA:At least one βi≠0
The detection methods for multicollinearity are mostly informal. Which of the following indicate a potential multicollinearity issue? Choose all that apply! Multiple choice question. Significant F statistic coupled with individually insignificant predictor variables Individually insignificant predictor variables High R2 and significant F statistic coupled with insignificant predictor variables High R2 plus individually insignificant predictor variables
High R2 and significant F statistic coupled with insignificant predictor variables
Consider the following linear regression model, which links the response variable y with k predictor variables x1, x2,..., xk:y=β0+β1x1+β2x2+... +βkxk+ε. If, for example, the slope coefficient β1 equals zero, then the predictor variable x1 does what and implies what? Choose all that answer the 'does what?' and 'Implies what?' questions! Multiple select question. Implying that x1 does not influence y Then x1 influences y Does not drop out of the equation Drops out of the equation
Implying that x1 does not influence y Drops out of the equation
A simple linear regression model and is represented as y = β0 + β1x1 + ɛ,; What do β0and β1 (the Greek letters read as betas) represent? (They must be shown in the correct order!) Multiple choice question. Slope, dependent Dependent, independent Slope, intercept Intercept, slope
Intercept, slope
The nonzero slope coefficient test is used for a renowned financial application referred to as the capital asset pricing model (CAPM).The model y = α + βx + ɛ, is essentially a simple linear regression model that uses α and β, in place of the usual β0 and β1, to represent the intercept and the slope coefficients, respectively. Which of the following is true about the slope coefficient β, called the stock's beta? Select that apply! Multiple select question. Measures how sensitive the stock's return is to changes in the level of the overall market A stock for which β > 1 is considered more "aggressive" or riskier than the market When β equals 0, any change in the market return leads to an identical change in the given stock return. When β equals 1, any change in the market return leads to an identical change in the given stock return.
Measures how sensitive the stock's return is to changes in the level of the overall market When β equals 1, any change in the market return leads to an identical change in the given stock return. A stock for which β > 1 is considered more "aggressive" or riskier than the market
What is a good solution when confronted with multicollinearity? Multiple select question. Obtain more data because the sample correlation may get weaker Drop one of the collinear variables Obtain more data because a bigger sample is always better Add another variable
Obtain more data because the sample correlation may get weaker Drop one of the collinear variables
The variance inflation factor (VIF) is another measure that can detect a high correlation between three or more predictor variables even if no pair of predictor variables has a particularly high correlation. What is the smallest possible value of VIF? (absence of multicollinearity). Multiple choice question. Zero VIF exceeds 5 or 10 VIF does not exceed 5 or 10 One
One
What is the condition called when two or more predictor variables have an exact linear relationship? Multiple choice question. Model inadequacies Nonzero slope coefficient Perfect multicollinearity Nonlinear violation
Perfect multicollinearity
If a linear regression model uses only one predictor variable, then the model is referred to as a 'BLANK' linear regression model
Simple
In the presence of changing variability, the estimated standard errors of the OLS estimators are inappropriate. What does this imply about using standard testing? Multiple choice question. We should use F tests only Use standard t or F tests We should use standard t tests only Standard t or F tests are not valid as they are based on these estimated standard errors.
Standard t or F tests are not valid as they are based on these estimated standard errors.
When we assess a linear regression model, there are several tests we can use. What is the test called that determines whether the predictor variables x1, x2,..., xk have a joint statistical influence on y? Multiple choice question. Test of individual significance Test for nonzero slope coefficient Test of joint significance Test for goodness of fit
Test of joint significance
What is used to evaluate how well the sample regression equation fits the data? Multiple choice question. The dispersion of residuals The goodness-of-fit measure The coefficient of determination, R² The standard error of the estimate
The coefficient of determination, R²
We can use residual plots to gauge changing variability.The residuals are generally plotted against each predictor variable xj Which of the following indicates there is no violation? Multiple choice question. There is no way to indicate no violation The predictor variable is randomly dispersed across the residuals The residuals are NOT randomly dispersed across the values of xj The residuals are randomly dispersed across the values of xj
The residuals are randomly dispersed across the values of xj
We can plot the residuals sequentially over time to look for correlated observations. If there is no violation, then what would you see? Multiple choice question. The residuals should show a normal pattern around the vertical axis. The residuals should show a normal pattern around the horizontal axis. The residuals should show no pattern around the horizontal axis. The residuals should show no pattern around the vertical axis.
The residuals should show no pattern around the horizontal axis.
In order to avoid the possibility of R2 creating a false impression, virtually all software packages include adjusted R2. Unlike R2, adjusted R2 explicitly accounts for what? Multiple select question. The sample size n The number of predictor variables k The number of samples taken Multicollinearity
The sample size n The number of predictor variables k
Instead of se2,we generally report the standard deviation of the residual, denoted se, more commonly referred to as Multiple choice question. The descriptive statistic Goodness-of-fit The standard error of the estimate The standard deviation of the sample
The standard error of the estimate
In order to select the preferred model, we examine several goodness-of-fit measures: Select all goodness-of-fit measures examined! Multiple select question. The standard error of the estimate The coefficient of determination The adjusted coefficient of determination The standard coefficient
The standard error of the estimate The coefficient of determination The adjusted coefficient of determination
We use analysis of variance (ANOVA) in the context of the linear regression model to derive R2.We denote the total variation in y as Σ(yi−y ̄)2, which is the numerator in the formula for the variance of y. What is this total variation called? Multiple choice question. Total sum of squares Regression error Total error Squared error
Total sum of squares
True or false: Linearity is justified if the residuals are randomly dispersed across the values of a predictor variable. True false question. True False
True
An important first step before running a regression model is to compile a comprehensive list of potential predictor variables. How can we reduce the list to a smaller list of predictor variables? Multiple choice question. The best approach may be to do nothing We must include all relevant variables We use R to make the necessary correction Use the adjusted R2 criterion to reduce the list
Use the adjusted R2 criterion to reduce the list
If one or more of the relevant predictor variables are excluded, then the resulting OLS estimators are biased. The extent of the bias depends on the degree of the 'BLANK" between the included and the excluded predictor variables.
correlation
When comparing models with the same response variable, we prefer the model with a smaller se. A smaller se implies that there is 'BLANK' dispersion of the observed values from the predicted values.
less
If we include as many dummy variables as there are categories, then their sum will equal 'BLANK'.
one called dummy variable trap
'BLANK' plots are used to detect some of the common violations to the regression model assumptions. These graphical plots are easy to use and provide informal analysis of the estimated regression models.
residual
If residual plots exhibit strong nonlinear patterns, the inferences made by a linear regression model can be quite misleading. In such instances, we should employ nonlinear regression methods based on simple transformations of the 'BLANK' and the predictor variables.
response
Suppose the competing hypotheses in testing for individual significance are H0: βj = 0 versus HA: βj ≠ 0. What would rejecting the null hypothesis imply? Multiple choice question. xj is significant in explaining y xj is not significant in explaining y xjexplains all the variation in y We would accept the null hypothesis
xj is significant in explaining y
The simple linear regression model y = β0 + β1x + ɛ implies that if x goes up by one unit, we expect y to change by how much? (irrespective of the value of x), Multiple choice question. x β1 β0 ɛ
β1