research methods exam 2

¡Supera tus tareas y exámenes ahora con Quizwiz!

Three types of operationalization

self-report observational physiological

Todd is studying the effect of popularity on academic success for his research methods project. He decides to measure popularity by asking each elementary school student to tell him how many friends he or she has. He assumes that more friends means the student is more popular. Which of the following best describes this variable? A ratio scale of measurement A qualitative variable An other-report measure A categorical variable

A

Which of the following is NOT an empirical test of validity? a) Face validity b) Criterion validity c) Divergent validity d) Convergent validity

A

What do you think might be some advantages and disadvantages of using a forced-choice format?

Advantage: Easy to analyze. Prevents fence-sitting (a problem with Likert scales we will address later). Disadvantage: Oversimplifies. Does not distinguish between strong and weak opinions, or degrees of confidence.

What do you think might be some advantages and disadvantages of using open-ended questions on a survey?

Advantage: People can respond however they want. You might learn something unexpected. Disadvantage: Can be difficult or impossible to score/analyze using statistical methods.

various ways to encourage honest responses on surveys.

Anonymous data collection Special "catch" questions Ask people's friends to rate them. Implicit tests

For his research methods class, Felipe plans to watch how students treat other children in their classrooms who have attention-deficit/hyperactivity disorder (ADHD). He will evaluate how positively or negatively the children are treated by their classmates. This is an example of what type of measurement? Self-report measurement Observational measurement Physiological measurement Archival measurement

B

To operationalize driving ability, a researcher counts the number of cones participants knock over on a driving course. What type of operational definition is this? Physiological Observational Self-report Objective

B

Which of the following is true of operational definitions? There is only one operational definition that is possible for each conceptual definition. The specification of operational definitions is one of the creative aspects of the research process. Conceptual definitions are created after operational definitions are determined. Operational definitions and conceptual definitions are the same thing

B

Why are convergent and discriminant validity often evaluated together? Incorrect Response Both terms mean the same thing. Both involve collecting information from a lot of psychological measures of theoretical interest. Both require the use of scatterplots. Both have to be determined by a panel of experts.

B

A physiological measure operationalizes a variable by recording a participants' a) Observable behaviors b) Intrapersonal thoughts c) Biological data d) Questionnaire answers

C

Another word for discriminant validity is ______________ validity. convergent asymmetrical divergent multiple

C

Which of the following is NOT possible? A measure is neither reliable nor valid. A measure is both valid and reliable. A measure is reliable but not valid. A measure is valid but not reliable.

D.

Imagine the following headline "Today's college students are more narcissistic than ever before". How would you question this study's external validity?

How did they sample the college students?

Imagine the following headline "Today's college students are more narcissistic than ever before". How would you question this study's construct validity?

How was narcissism measured (and how was it measured in the past?)

How would you find out if the two operational definitions of happiness measure the same construct?

One method: test if the results from the questionnaire and Ladder of Life are strongly correlated.

What's the difference between an interval and a ratio scale?

an interval scale has no true zero

How generalizable are the results of a study?

external validity

Research that attempts to reach every member of the population.

census

solutions to reactivity

blend in, wait it out, measure the behavior's results (instead of behavior directly)

example of observer effects:

clever hans

Select a random set of clusters from within a population, then sample everyone within those clusters.

cluster sampling

an abstract or theoretical variable

construct

describe the strength of an association, and can also be handy for evaluating reliability.

corelation coeficients (r)

For example: Should we cut funding for Art and Gym? yes? no?

double-barreled question

in this type of validity, your measure is correlated with a relevant outcome: how well the measure predicts a relevant outcome. How well does a test score predict on-the-job performance?

criterion validity

what are three empirical ways to assess validity?

criterion, convergent, discriminant validity

what are two subjective ways to assess validity?

face & content validity

Example: "Which do you prefer, Coke or Pepsi?" what type of question is this?

forced-choice format

Example: "What are your comments on this class?" what type of question is this?

open-ended

how the researcher measures the construct

operational definition

variables representing a rank order. Example: conference standings in sports

ordinal

a sample with a similar composition to the larger population.

representative sample

Only _________ allow us to make valid inferences about the population of interest.

representative samples

A smaller set of people that is supposed to represent the population.

sample

What does it mean to say a sample is biased?

the traits of the sample differ from the traits of the population

ways in which participants don't always give accurate responses to surveys:

using shortcuts (yea-saying) trying to look good self-reporting "more than they can know" false memories

Answering "strongly agree" to every question on a survey to finish more quickly. (acquiescence, or yea-saying) Or, Answering in the middle of the scale every time to avoid expressing controversial opinions (fence-sitting) how to discourage this?

yea-saying (using shortcuts) reverse-coding, eliminate the middle option

Variables whose values are numbers.

quantitative variables

External validity is extremely important for evaluating ________claims.

frequency

If you ask multiple people to observe the same thing, do they reach the same results? which type of reliability?

interrater

"Do you have any problems with your boss?" "How much do you love eating delicious Papa John's pizza?"

leading questions

A correlation-based statistic called _____________ is commonly used to determine internal reliability. Cronbach's alpha kappa a scatterplot Pearson's r

A

a sample whose composition differs from the larger population. This can occur when certain members of the population are more likely to be sampled than others.

biased sample

Select a random set of clusters from within a population, then sample a subset of people within those clusters.

multistage sampling

What is divergent validity?

the idea that measures of unrelated constructs should not be strongly associated

"Does it seem possible or does it seem impossible to you that the Nazi extermination of the Jews never happened?"

negatively-worded questions

Variables whose values are different categories.

categorical variables

in this type of validity, your measure is less strongly associated with measures of dissimilar constructs Does the measure correlate less strongly with measures of different constructs?

discriminant validity

this type of validity looks like what you want to measure Does it look like a good measure? (often assessed by asking experts)

face validity

In one study, 73% of students surveyed believed they could recall seeing the first plane hit the tower on 9/11, even though no such footage was shown at the time.

false memories

If you ask the same question in different ways, do you get the same answer? what type of reliability?

internal

Equal intervals between values, but a value of zero doesn't mean "nothing". Example: IQ test

interval

Some things to avoid when writing a well-worded question

leading questions, double-barreled questions, negatively-worded questions, question order effects

People do not always know the true reasons for their preferences (Nisbett & Wilson, 1977).

self-reporting more than they can know

How many subcategories of quantitative variables exist? Two Three Four Five

three

is used more often when it's about social/political opinions.

poll

What are the three ethical principles of the Belmont Report?

respect for persons, beneficence, justice

A sample is biased when

the composition of the sample differs from the population of interest.

What would you do if you witnessed someone being mugged in the street? how would people respond to this by using a shortcut?

trying to look good

randomly assigning your participants to different groups/conditions.

random assignment

randomly selecting the participants for your research study.

random sampling

Equal intervals between values, and a value of zero really means "nothing". Example: Reaction times

ratio

When people change their behavior if they know they are being watched.

reactivity

refers to the appropriateness of an operational definition. Does the test measure what it's supposed to?

construct validity

in this type of validity, the measure contains all the parts that your theory says it should contain: Does it include all the important contents of the construct?

content validity

Sometimes a researcher is not overly concerned with external validity. In these cases, they may settle for biased sampling techniques.

convenience sampling, purposive sampling, snowball sampling, quota sampling

In this type of validity, your measure is more strongly associated with measures of similar constructs. Does the measure correlate with other measures of the same construct?

convergent validity

A number which depends on the average correlation among questions which are supposed to measure the same construct. A value of _____ or higher is generally considered good

cronbach's alpha .7

what are some ways of preventing observer bias and observer effects?

observers trained using coded rating scales, multiple observers (interrater reliability), coders observers not aware of the study's hypothesis

intentionally sampling more people from a group with a low population in order to get statistically meaningful results. The final results must be adjusted so the opinions of the oversampled group are weighted to their actual population.

oversampling

In another study, researchers observed the emotional tone and topics of conversation during family dinners. To code emotional tone, researchers rated parents on a scale from 1 (cold/hostile) to 7 (warm/happy). The average was slightly positive (4.2). In addition, kids talked more about how they didn't like the food, and parents talked more about how healthy it was. What type of reliability would be particularly important here?

interrater-reliability

Another way to evaluate criterion validity: check whether the measure picks up on known differences between groups Example: to test a lie detector, we would want to try it on people we know are lying, and people we know are telling the truth.

known-groups paradigm

Example: "I am able to do most things as well as other people" 2 3 4 5 6 7 strongly disagree strongly agree what type of question is this?

likert scale

The entire set of people that a researcher is interested in.

population

Hosea is studying the relationship between caffeine consumption and problem-solving ability. Which of the following is a categorical way to operationalize caffeine consumption? The number of cups of coffee consumed in a day The number of milligrams of caffeine consumed during the study The frequency of buying energy drinks Whether the participant drank a soda in the 24 hours prior to the study

D

What is a leading question?

a question that is framed in such a way as to encouraged a particular response

what are the three types of reliability?

test-retest interrater internal

Many claims are not meant to generalize to the entire human population. Instead, researchers first identify a _____________ (e.g. latino voters), and figure out how to sample it.

population of interest

(aka random sampling): every member of the population has an equal chance to be sampled.

population sampling

Example: Does your child play sports? How often does your child play?

question order effect

What are two things a researcher can do if he/she is concerned about acquiescence (yea-saying) on a survey?

reverse-code some of the questions ask special "catch" questions that nobody should strongly agree with if they're paying attention

ways to get a biased sample:

sampling only those who are convenient to contact, sampling only those who invite themselves (self-selection)

uses a random selection process. Every member of the population of interest could potentially be chosen. Usually not realistic.

simple random sampling

different ways to do population sampling:

simple random sampling, cluster & multistage samling, stratified random sampling, oversampling, systematic sampling

Establishing construct validity would probably be most important for which of the following? A measure of heart rate A measure of the number of times a person eats alone during a month A measure of spirituality A measure of income

spirituality

Select a variety of groups non-randomly, but randomly sample within those groups

stratified random sampling

is used more often when it relates to opinions about consumer products.

survey

methods of posing questions to people.

surveys and polls

using a systematic (but arbitrary) method to sample participants.

systematic sampling

do you get the same results when you test someone twice? Which type of reliability

test-retest

When a researcher watches people or animals and records what they do.

observational research

Two groups of psychoanalysts were shown the same interview of a 26-year-old man discussing his feelings and work experiences. One group was told the man was a patient. The other group was told he was a job applicant. When asked to describe him, the two groups described the man using totally different adjectives: "innovative" vs. "afraid of his own aggressive impulses". example of?

observer bias

when an observer's expectations influence their interpretation of behavior.

observer bias

what are some issues related to construct validity to avoid:

observer bias, observer effects, reactivity

Similar to observer bias. However, instead of just being biased in their interpretations of behavior, the researcher actually changes the behavior of those they are observing.

observer effects

you could test your participants at two different times, and compute the correlation between scores at time 1 and time 2. IQ is supposed to be stable over time, so you would expect a strong, positive correlation. what is this testing? (which reliability?)

test-retest

The popular press often reports on parents acting violently at their children's hockey games. To estimate the true frequency of such behavior, researchers observed 69 boys and girls hockey games in one Canadian city. They observed 64% positive comments, only 4% negative. How would you question this study's external validity? How could one test the convergent validity of this study?

was the one canadian city they studied a typical example? (generalizability) repeat the study using a different measure of violent behavior


Conjuntos de estudio relacionados

GEB6895 Federal Reserve and Monetary Policy

View Set

EAQ 6: Coping and Stress Tolerance

View Set

Chapter 1: Database Systems, Chapter 2: Data Models, Chapter 3: The Relation Database Model, Chapter 4: Entity Relationship ER Modeling, Chapter 6: Normalization of Database Tables, Chapter 7: Introduction to Structured Query Language (SQL), Chapter...

View Set

Primerica - Life Insurance Policy Provisions, Options & Riders (AZ)

View Set