Research Methods
Content Analysis
"A technique for systematically describing written, spoken or visual communication. It provides a quantitative (numerical) description." Reading a text using a coding sheet (created via a theory); Give the content a nominal scale for measure Can't be used to test causality; cannot determine if something is the cause of something else Can determine prevalence of a problem/ issue/ phenomenon
Critical Scholarship: Ontology
All reality is socially constructed (radical constructivist will believe there is nothing but what we have socially constructed)
Social Science: Assumptions
Answers to our questions mean something
Content Validity
Assessment of how well a set of scale items matches with the relevant content domain of the construct that it is trying to measure.
Critical Scholarship: Methodology
Assumption dependent
Social Science: Ontology
Believes in a reality that exist independently to
Causality & how to demonstrate it
Causality: Whether the observed change in dependent variable is indeed caused by a corresponding change in hypothesized independent variable and not variables extraneous to the research context. Causality requires three conditions: (1) covariation of cause and effect (i.e., if cause happens, then effect also happens; and if cause does not happen, effect does not happen), (2) temporal precedence: cause must precede effect in time, (3) no plausible alternative explanation (or spurious correlation).
Critical Scholarship: Epistemology
Considering varying individual perspectives
Inductive Reasoning
Empirical (specific) to Theoretical (general)
Empirical Assessment of Validity
Empirical assessment of validity examines how well a given measure relates to one or more external criterion, based on empirical observations.
Social Science:Epistemology
Empirical observation; we can observe in a post positivist (logical assumptions) way
Internal Validity
Evidence that any observed outcomes in the dependent variable can conclusively be attributed to the manipulation and not to some extraneous variable or source. "Did the manipulation really cause the outcome?"
Concurrent Validity
How well one measure relates to other concrete criterion that is presumed to occur simultaneously
Manipulation Check
Important task in experimental design: Checking for the adequacy of design. Pilot tests, pre tests, post tests all serve as manipulation checks. / A means of assessing whether or not the manipulation worked as intended - E.g. Manipulating single player vs. multiplayer games; you can ask after playing a multiplayer game if the children cooperated with others or played alone.
Stratified sampling
In stratified sampling, the sampling frame is divided into homogeneous and non-overlapping subgroups (called "strata"), and a simple random sample is drawn within each subgroup. (Bhatta pg. 67)
Pretest-posttest control group design
In this design, subjects are randomly assigned to treatment and control groups, subjected to an initial (pretest) measurement of the dependent variables of interest, the treatment group is administered a treatment (representing the independent variable of interest), and the dependent variables measured again (posttest). (Bhatta pg. 86)
Quota sampling
In this technique, the population is segmented into mutually- exclusive subgroups (just as in stratified sampling), and then a non-random set of observations is chosen from each subgroup to meet a predefined quota (bhatta pg. 69)
Variable
Measurable representation of a construct
Negative Relationship
Negative correlation is a relationship between two variables in which one variable increases as the other decreases, and vice versa.
Paradigm
Our worldview; frame that influences how and what we think
Sampling
Sampling is the statistical process of selecting a subset (called a "sample") of a population of interest for purposes of making observations and statistical inferences about that population.
Operational Definitions
Scientific research requires operational definitions that define constructs in terms of how they will be empirically measured.
Convergent Validity
The closeness with which a measure relates to (or converges on) the construct that it is purported to measure
Discriminant Validity
The degree to which a measure does not measure (or discriminates from) other constructs that it is not supposed to measure.
Deductive Reasoning
Theoretical (general) to empirical (specific)
Theoretical Assessment of Validity
Theoretical assessment of validity focuses on how well the idea of a theoretical construct is translated into or represented in an operational measure.
Posttest-only control group design
This design is a simpler version of the pretest- posttest design where pretest measurements are omitted (Bhatta pg. 86)
Attrition
Threat to Validity: when participants drop out of the study before completion (various reasons for this)
Type 1 Error + Type 2 Error
Type 1: Getting a false positive to the hypothesis. They way you collect data can result in a Type 1 or Type 2 error. Type 2: Getting a false "no relationship" even if there is.
Validity
Validity, often called construct validity, refers to the extent to which a measure adequately represents the underlying construct that it is supposed to measure. For instance, is a measure of compassion really measuring compassion, and not measuring a different construct such as empathy?
Hypothesis
The empirical formulation of propositions, stated as relationships between variables, is called hypotheses. Propositions are specified in the theoretical plane, while hypotheses are specified in the empirical plane. Hence, hypotheses are empirically testable using observed data, and may be rejected if not supported by empirical observations.
Praxis
The idea that scholarship should serve to aid in the emancipation of people from power
Operational Linkages
The predicted connection or relationship between the operationalized variables in the hypothesis, first, and in the statistical analyses, second, is referred to as the operational linkage. Operational linkages are those link that predict and test theoretical linkages. Krcmar, pg. 11.
Theory and its relationship to conceptual model
The process of theory or model development may involve inductive and deductive reasoning. (*Your theory will inform whether you should be using inductive or deductive reasoning to form your conceptual model). Deduction is the process of drawing conclusions about a phenomenon or behavior based on theoretical or logical reasons and an initial set of premises. As an example, if a certain bank enforces a strict code of ethics for its employees (Premise 1) and Jamie is an employee at that bank (Premise 2), then Jamie can be trusted to follow ethical practices (Conclusion). In deduction, the conclusions must be true if the initial premises and reasons are correct. In contrast, induction is the process of drawing conclusions based on facts or observed evidence. For instance, if a firm spent a lot of money on a promotional campaign (Observation 1), but the sales did not increase (Observation 2), then possibly the promotion campaign was poorly executed (Conclusion). As shown in Figure 2.3, inductive and deductive reasoning go hand in hand in theory and model building.
Theoretical Linkages
Theories provide explanations of social or natural phenomenon. Theoretical linkages are propositional statements of relationship stating that two variables are linked in such a way that they are either associatively or causally related. Ex: "As exposure to television violence increases, aggression levels increase" OR "Exposure to television violence causes increases in aggressive behavior" Krcmar, pg. 7
Switched Replication Design
This is a two-group design implemented in two phases with three waves of measurement. The treatment group in the first phase serves as the control group in the second phase, and the control group in the first phase becomes the treatment group in the second phase, as illustrated in Figure 10.7. In other words, the original design is repeated or replicated temporally with treatment/control roles switched between the two groups. By the end of the study, all participants will have received the treatment either during the first or the second phase. This design is most feasible in organizational contexts where organizational programs (e.g., employee training) are implemented in a phased manner or are repeated at regular intervals. (bhatta pg. 89)
Randomized block design
This is a variation of the posttest-only or pretest-posttest control group design where the subject population can be grouped into relatively homogeneous subgroups (called blocks) within which the experiment is replicated. For instance, if you want to replicate the same posttest-only design among university students and full-time working professionals (two homogeneous blocks), subjects in both blocks are randomly split between treatment group (receiving the same treatment) or control group
Sampling Frame
This is an accessible section of the target population (usually a list with contact information) from where a sample can be drawn. If your target population is professional employees at work, because you cannot access all professional employees around the world, a more realistic sampling frame will be employee lists of one or two local companies that are willing to participate in your study.
Statistical Regression to the Mean
Threat to Validity: phenomenon that can be particularly problematic when researchers use a repeated measure design in which participants are measured on a given variable more than once. Those who score really high or really low tend to move toward the sample mean on subsequent measures.
Selection Bias
Threat to validity: When participants self select into a study, researchers must determine if those who are in the study are similar to the population of interest and whether they behave in unusual ways simply because they are willing to participate.
Hawthorne Effect
Threat to validity: when outcomes or changes in the participants occur because they know they are being observed. Wanting to perform to please the experimenters.
Axiological
Value assumption: study and the nature or value and valuation, the kinds of things that are valuable- What values drive what we know? Value-free = objectivism / Value-intended = subjectivism.
Mediating variables
Variables that are explained by independent variables while also explaining dependent variables are mediating variables (or intermediate variables)
Dependent Variable
Variables that are explained by other variables are dependent variables
Independent Variable
Variables that explain other variables are called independent variables
Moderating Variables
Variables that influence the relationship between independent and dependent variables are called moderating variables.
Ontology Definition
What we think the world is- e.g. subjective, objective.. (our paradigm influences our ontology)
Face Validity
Whether an indicator seems to be a reasonable measure of its underlying construct "on its face". Ex: frequency of one's attendance at a religious service seems to make sense as an indicator of a person's religiosity without a lot of explanation.
Systematic sampling
the sampling frame is ordered according to some criteria and elements are selected at regular intervals through that ordered list. Systematic sampling involves a random start and then proceeds with the selection of every kth element from that point onwards, where k = N/n, where k is the ratio of sampling frame size N and the desired sample size n, and is formally called the sampling ratio. (Bhatta pg. 67)
Implicit Theory
System 1 theories: "our perceptions, interpretations, understanding, and even memory and recall occur at the preconscious level.
Explicit Theory
System 2 theories: "when we read, draw conclusions, and perhaps summarize what we have read in a systematic, consciously aware way."
Positivism
(Auguste Comte) At its core, positivism is based on the assumption of an observable, law-based reality. Knowledge is limited to claims about the world that are unambiguously (positively) true, either because they are based on verifiable observations or because they must logically follow from such an observation.
Post-Positivism
(Karl Popper) (Falsification) We cannot be certain that what we think is true, but we can be certain about what we know not to be true.A theoretical model could be valid if one fails to observe empirical evidence inconsistent with it or even contradicting it when trying to falsify the theoretical model.
Systematic Error
(Measurement Error) Systematic error is an error that is introduced by factors that systematically affect all observations of a construct across an entire sample in a systematic manner. In our previous example of firm performance, since the recent financial crisis impacted the performance of financial firms disproportionately more than any other type of firms such as manufacturing or service firms, if our sample consisted only of financial firms, we may expect a systematic reduction in performance of all firms in our sample due to the financial crisis. Unlike random error, which may be positive negative, or zero, across observation in a sample, systematic errors tends to be consistently positive or negative across the entire sample.
Random Error
(Measurement error) The error that can be attributed to a set of unknown and uncontrollable external factors that randomly influence some observations but not others. As an example, during the time of measurement, some respondents may be in a nicer mood than others, which may influence how they respond to the measurement items.
Conceptual Model
A conceptual model is a representation of a system, made of the composition of concepts which are used to help people know, understand, or simulate a subject the model represents
Constructs
A construct is an abstract concept that is specifically chosen (or "created") to explain a given phenomenon. A construct may be a simple concept, such as a person's weight, or a combination of a set of related concepts such as a person's communication skill, which may consist of several underlying concepts such as the person's vocabulary, syntax, and spelling. The former instance (weight) is a unidimensional construct, while the latter (communication skill) is a multi-dimensional construct (i.e., it consists of multiple underlying concepts).
Content Analysis
A content analysis is used to analyze the relationship between the measurement rules used to assign numerical value to messages in order to describe the messages and make inferences about the meaning.
Model
A model is a representation of all or part of a system that is constructed to study that system (e.g., how the system works or what triggers the system). While a theory tries to explain a phenomenon, a model tries to represent a phenomenon. Models are often used by decision makers to make important decisions based on a given set of inputs. For instance, marketing managers may use models to decide how much money to spend on advertising for different product lines based on parameters such as prior year's advertising expenses, sales, market growth, and competing products. Likewise, weather forecasters can use models to predict future weather patterns based on parameters such as wind speeds, wind direction, temperature, and humidity. While these models are useful, they may not necessarily explain advertising expenditure or weather forecasts. Models may be of different kinds, such as mathematical models, network models, and path models. Models can also be descriptive, predictive, or normative.
Population
All people or items (unit of analysis) with the characteristics that one wishes to study - If you wish to identify the primary drivers of academic learning among high school students, then what is your target population: high school students, their teachers, school principals, or parents? The right answer in this case is high school students, because you are interested in their performance, not the performance of their teachers, parents, or schools.
Convenience sampling
Also called accidental or opportunity sampling, this is a technique in which a sample is drawn from that part of the population that is close to hand, readily available, or convenient. (Bhatta pg. 69)
Threats to Validity
Anything that happens during an experiment that can question if the manipulation really caused the outcome.
Positivism
Argues that for something to be valid it must be observable (see it to believe it); anti metaphysical- can't touch it, it's not real
Theory
Broad explanations of/ predictions of behavior or phenomena. / A theory is a set of systematically interrelated constructs and propositions intended to explain and predict a phenomenon or behavior of interest, within certain boundary conditions and assumptions. Essentially, a theory is a systematic collection of related theoretical propositions. While propositions generally connect two or three constructs, theories represent a system of multiple constructs and propositions.
Characteristics of a true experiment
One or more independent variables are manipulated by the researcher (as treatments), subjects are randomly assigned to different treatment levels (random assignment), and the results of the treatments on outcomes (dependent variables) are observed. (Types of Experiments: Pretest-posttest control group design, posttest only control group design, etc).
Test-retest reliability.
Test-retest reliability is a measure of consistency between two measurements (tests) of the same construct administered to the same sample at two different points in time. If the observations have not changed substantially between the two tests, then the measure is reliable.
Predictive Validity
The degree to which a measure successfully predicts a future outcome that it is theoretically expected to predict
Split-half reliability
Split-half reliability is a measure of consistency between two halves of a construct measure. For instance, if you have a ten-item measure of a given construct, randomly split those ten items into two sets of five (unequal halves are allowed if the total number of items is odd), and administer the entire instrument to a sample of respondents. Then, calculate the total score for each half for each respondent, and the correlation between the total scores in each half is a measure of split-half reliability.
Multi-Stage Sampling
Depending on your sampling needs, you may combine these single-stage techniques to conduct multi-stage sampling. For instance, you can stratify a list of businesses based on firm size, and then conduct systematic sampling within each stratum. This is a two-stage combination of stratified and systematic sampling. Likewise, you can start with a cluster of school districts in the state of New York, and within each cluster, select a simple random sample of schools; within each school, select a simple random sample of grade levels; and within each grade level, select a simple random sample of students for study. In this case, you have a four-stage sampling process consisting of cluster and simple random sampling. (Bhatta pg. 68)
Measurement Validity (psychometric theory)
Examines how measurement works, what it measures, and what it does not measure. (Bhatta pg. 55) Scales indeed measure the unobservable construct that we wanted to measure (scales are valid) and they measure the intended construct consistently and precisely (reliable).
Criterion-related Validity
Examines whether a given measure behaves the way it should, given the theory of that construct.
Social Science: Methodology
Experiments, content analysis, surveys, ethnographies, interviews
Concepts
Explanations require development of concepts or generalizable properties or characteristics associated with objects, events, or people. While objects such as a person, a firm, or a car are not concepts, their specific characteristics or behavior such as a person's attitude toward immigrants, a firm's capacity for innovation, and a car's weight can be viewed as concepts.
Cluster Sampling
If you have a population dispersed over a wide geographic region, it may not be feasible to conduct a simple random sampling of the entire population. In such case, it may be reasonable to divide the population into "clusters" (usually along geographic boundaries), randomly sample a few clusters, and measure all units within that cluster. (Bhatta pg. 68)
Solomon four-group design
In this design, the sample is divided into two treatment groups and two control groups. One treatment group and one control group receive the pretest, and the other two groups do not. This design represents a combination of posttest-only and pretest-posttest control group design, and is intended to test for the potential biasing effect of pretest measurement on posttest measures that tends to occur in pretest-posttest designs but not in posttest only designs. (Bhatta pg. 88)
Simple Random Sampling
In this technique, all possible subsets of a population (more accurately, of a sampling frame) are given an equal probability of being selected. Simple random sampling involves randomly selecting respondents from a sampling frame, but with large sampling frames, usually a table of random numbers or a computerized random number generator is used.
Inter-rater reliability
Inter-rater reliability, also called inter-observer reliability, is a measure of consistency between two or more independent raters (observers) of the same construct.
Internal consistency reliability
Internal consistency reliability is a measure of consistency between different items of the same construct. If a multiple-item construct measure is administered to respondents, the extent to which respondents rate those items in a similar manner is a reflection of internal consistency. This reliability can be estimated in terms of average inter-item correlation, average item-to-total correlation, or more commonly, Cronbach's alpha. As an example, if you have a scale with six items, you will have fifteen different item pairings, and fifteen correlations between these six items.
Social Science: Topics
Interpersonal, organizational, intercultural, etc.
Interval Scales
Interval scales are those where the values measured are not only rank-ordered, but are also equidistant from adjacent attributes. For example, the temperature scale (in Fahrenheit or Celsius), where the difference between 30 and 40 degree Fahrenheit is the same as that between 80 and 90 degree Fahrenheit. Likewise, if you have a scale that asks respondents' annual income using the following attributes (ranges): $0 to 10,000, $10,000 to 20,000, $20,000 to 30,000, and so forth, this is also an interval scale, because the mid-point of each range (i.e., $5,000, $15,000, $25,000, etc.) are equidistant from each other.
Interview Surveys
Interviews are a more personalized form of data collection method than questionnaires, and are conducted by trained interviewers using the same research protocol as questionnaire surveys (i.e., a standardized set of questions). However, unlike a questionnaire, the interview script may contain special instructions for the interviewer that is not seen by respondents, and may include space for the interviewer to record personal observations and comments. In addition, unlike mail surveys, the interviewer has the opportunity to clarify any issues raised by the respondent or ask probing or follow-up questions. (face-to-face, focus groups, telephone interviews)
Post Positivism
No objective reality; you can make logical assumptions without physical observation
How are Quasi-Experimental Designs different from a true experimental design?
No random assignment
Nominal Scales
Nominal scales merely offer names or labels for different attribute values. Nominal scales, also called categorical scales, measure categorical data. These scales are used for variables or indicators that have mutually exclusive attributes. Examples include gender (two values: male or female), industry type (manufacturing, financial, agriculture, etc.), and religious affiliation (Christian, Muslim, Jew, etc.). Even if we assign unique numbers to each value, for instance 1 for male and 2 for female, the numbers don't really mean anything (i.e., 1 is not less than or half of 2) and could have been easily been represented non-numerically, such as M for male and F for female.
Non -probability sampling
Nonprobability samplingis a sampling technique in which some units of the population have zero chance of selection or where the probability of selection cannot be accurately determined. (Bhatta pg. 69)
Empirical
Observable
Ordinal Scales
Ordinal scales are those that measure rank-ordered data, such as the ranking of students in a class as first, second, third, and so forth, based on their grade point average or test scores. However, the actual or relative values of attributes or difference in attribute values cannot be assessed. For instance, ranking of students in class says nothing about the actual GPA or test scores of the students, or how they well performed relative to one another. A classic example in the natural sciences is Moh's scale of mineral hardness, which characterizes the hardness of various minerals by their ability to scratch other minerals. For instance, diamonds can scratch all other naturally occurring minerals on earth, and hence diamond is the "hardest" mineral. However, the scale does not indicate the actual hardness of these minerals or even provides a relative assessment of their hardness. Ordinal scales can also use attribute labels (anchors) such as "bad", "medium", and "good", or "strongly dissatisfied", "somewhat dissatisfied", "neutral", or "somewhat satisfied", and "strongly satisfied".
Philosophy and its relationship to theory
Our mental modes (paradigms) impact/constrain our thinking abd reasoning about observed phenomenon. (Positivism/ Post Positivism/ Ontology/ Epistemology) Our ways of looking at the world impact how we study the world (objectively, subjectively) and therefore impact the theories we believe or develop.
Epistemology Definition
Our understanding of how to study the world; how we know what we know
Positive Relationship
Positive correlation is a relationship between two variables in which both variables move in tandem. A positive correlation exists when one variable decreases as the other variable decreases, or one variable increases while the other increases.
Probability Sampling
Probability sampling is a technique in which every unit in the population has a chance (non-zero probability) of being selected in the sample, and this chance can be accurately determined. Sample statistics thus produced, such as sample mean or standard deviation, are unbiased estimates of population parameters, as long as the sampled units are weighted according to their probability of selection.
Ratio Scales
Ratio scales are those that have all the qualities of nominal, ordinal, and interval scales, and in addition, also have a "true zero" point (where the value zero implies lack or non- availability of the underlying construct). Most measurement in the natural sciences and engineering, such as mass, incline of a plane, and electric charge, employ ratio scales, as are some social science variables such as age, tenure in an organization, and firm size (measured as employee count or gross revenues). For example, a firm of size zero means that it has no employees or revenues. The Kelvin temperature scale is also a ratio scale, in contrast to the Fahrenheit or Celsius scales, because the zero point on this scale (equaling -273.15 degree Celsius) is not an arbitrary value but represents a state where the particles of matter at this temperature have zero kinetic energy.
Rhetorical Ontology
Reality is socially constructed; communication of language can construct subjective reality for audience
Reliability
Reliability is the degree to which the measure of a construct is consistent or dependable. In other words, if we use this scale to measure the same construct multiple times, do we get pretty much the same result every time, assuming the underlying phenomenon is not changing? An example of an unreliable measurement is people guessing your weight.
4 components of Empirical Inquiry
Replicable, Precise (theoretical constructs must be concisely defined), Empirically Testable, Falsifiability (anything that can possibly be disproven- Faith is not falsifiable), Parsimonious (simple- not too many conditions)
Conceptual Model's relationship to study design
Research designs vary based on whether the researcher starts at observation and attempts to rationalize the observations (inductive research), or whether the researcher starts at an ex ante rationalization or a theory and attempts to validate the theory (deductive research).
Questionnaire Surveys
Research instrument consisting of a set of questions (items) intended to capture responses from respondents in a standardized manner. Questions may be unstructured or structured. Unstructured questions ask respondents to provide a response in their own words, while structured questions ask respondents to select an answer from a given set of choices.
Survey
Research method involving the use of standardized questionnaires or interviews to collect data about people and their preferences, thoughts, and behaviors in a systematic manner.
How to write a hypothesis
Should include: Directionality (whether it is positive or negative), causality (whether x causes y, or y causes x), should clearly identify independent and dependent variables, and should be able to be evaluated as either true or false.
Covariance designs
Sometimes, measures of dependent variables may be influenced by extraneous variables called covariates. Covariates are those variables that are not of central interest to an experimental study, but should nevertheless be controlled in an experimental design in order to eliminate their potential effect on the dependent variable and therefore allow for a more accurate detection of the effects of the independent variables of interest. The experimental designs discussed earlier did not control for such covariates. A covariance design (also called a concomitant variable design) is a special type of pretest posttest control group design where the pretest measure is essentially a measurement of the covariates of interest rather than that of the dependent variables. (Bhatta pg. 86)
Matched Pair Sampling
Sometimes, researchers may want to compare two subgroups within one population based on a specific criterion. For instance, why are some firms consistently more profitable than other firms? To conduct such a study, you would have to categorize a sampling frame of firms into "high profitable" firms and "low profitable firms" based on gross margins, earnings per share, or some other measure of profitability. You would then select a simple random sample of firms in one subgroup, and match each firm in this group with a firm in the second subgroup, based on its size, industry segment, and/or other matching criteria. (Bhatta pg. 68)
External Validity
Studies that can be readily generalized beyond the laboratory setting are considered to be externally valid. Studies that take place outside of the lab in more natural settings have to allow for a lower degree of control but more accurately represent the real world. Thus, they may be more externally valid.