OT Assessment Exam 1: Lecture Content

Réussis tes devoirs et examens dès maintenant avec Quizwiz!

How to control variables leading to error

-Be trained in administering and interpretation of test -Recheck that correct procedures are being used -Be aware of biases, natural tendencies -Be aware of the influences of ones background, presence, or expectations -Be unobtrusive, and aware of test takers variable that might apply -Document any bias

Norm Referenced Assessments

-Compare ratings of individual to a normative sample (above or below average) -Outcome: able to evaluate client in relation to group -Use of bell curve

How to use non-standards BEFORE assessment

-Consider potential biases -limit or control potential factors -select assessment relevant to construct being measured

Evaluation requires

-Consideration of occupational performance, client factors, performance skills/patterns, contextual and environmental factors, activity demands -Synthesis of all data obtained -analytical interpretation of data -reflective reasoning

Knowledge needed for effective evaluation

-Diagnosis and conditions -OT constructs -Professional guidelines -Assessments -Resources -Psychometrics

Descriptive Assessment

-Diagnostic -Measures a specific behavioral domain -Tells us whether there is a prob or not -Only given once

How to use non-standards AFTER assessment

-Evaluation findings -Synthesize info -Connect gathered info on theoretical foundation

Observation

-Includes but is not limited to: odors/aromas, temp/tone/color of skin, affect/mood, postures/gestures, speech patterns/eye contact, manner of dress/grooming, response to situation

Evaluation Provides us with?

-Info about pts needs and problems -Info about uniqueness of individual including contextual and environmental aspects -baseline for treatment to evaluate progress

Non-Standardized Assessments

-Ipsative Referenced -Dynamic -High internal validity -Highly individualized info -NOT generalized to other people

Static Assessments

-Measure ability (ex: grip strength) -No feedback given to client -No adjustments to instructions -One-way relationship (therapist >> client) -Results may or may not facilitate intervention planning

Dynamic Assessments

-Measure processes (ex: learning process) -Feedback given to client -Adjustments made in gradation -Two way relationship (therapist <<>>client) -Results should directly facilitate intervention planning

Criterion Referenced Tests in OT

-Measure response to intervention over time -School function assessment (SFA) -Klein-bell assessment of daily living skills -Kohlmann Evaluation of daily living skills (KELS) -Arnadottir OT-ADL Neurobehavioral Evaluation (A-ONE) -Performance Assessment of Self-Care skills (PASS)

Observation Examples

-Model of Human Occupation Screening Tool (MOHOST)

How to use non-standards DURING assessment

-Monitor personal biases/environmental influences -Monitor persons autonomic sys, energy level -Congruence of body language and actions

Types of Observation

-Naturalistic (natural environment) -Analogue (Simulated environment) -Participant Monitoring

Standardized Assessments

-Norm referenced -Criterion Referenced -Static -Have clear instructions for administration and scoring -Can be generalized to general population

Occupational Performance Assessment

-Primary method of data collection -requires activity analysis -often includes rating component

Interview

-Process of inquiry -Informal to formal -Open-ended or closed-ended questions -Info factors heavily into clinical reasoning

Questionnaire

-Self-report that can be completed with or without OT present -Provides info from clients perspective

Assessment

-Should address occupational performance and factors that support occupational performance -Selection should involve several components (from assessment critique)

Criterion-Referenced Approach

-Test results are interpreted by content mastery -Determines what they can do/know -NOT compared with others/average -Show presence or absence of characteristic, ability, or skill -all-or-none score (mastered or not mastered) -need to meet cut score/minimum to passed

Evaluative Assessment

-Useful for detecting clinical change -Grades behavior or performance -Pre-post -Most common

Predictive Assessment

-Useful for prognosis or outcomes -Compares performance/behavior against a criterion -Ex: Fall Screen

Correlation co-efficient > r

-Varies from -1.00 to +1.00 -strength determined by how close r -value is to +/- 1.00 -not a percentage (to find percentage square co-efficient)

Why choose non-standardized measure

-You want individualized perspective of a clients performance -Limited resources, easier to administer, inexpensive, takes less time, does not require alot of training

Concurrent Validity

-compare to gold-standard test

Reliability

-degree of consistency between two test administrations (stability) -Evaluator finds same result each time they conduct test

Validity

-degree to which a test accurately measures the specific construct trait, behavior, or performance it was designed to measure -is it measuring what it claims to measure (is it on target)

Content Validity

-descriptive not statistical -items represent a sufficient representative sample of domain/construct being examined -literature-supported

Predictive Validity

-makes predictions about future behaviors -extent to which scores on current test forecast measure of future criterion ex: fall risk screen

Test Re-test reliability

-measure of test score stability on the same version of the test repeated over two occasions -want a "not significant" result

Face Validity

-no statistical measure, w/o statistical proof -based on appearance -subjective, informal

Evaluation

-only OT may perform evaluation -Plan of care must be based on eval -must have face-to-face interaction w/ client during eval

Test Reliability

-reported using correlation statistic -examines consistency of results -test can be reliable but not valid

Screening

-used to determine need for eval -not reimbursable typically -used to generate referral for OT -Short time period (15 min, 1 unit of service) -Must comply with current state laws and regulatory requirements -may be preformed by OT or COTA (in texas)

Error: Logical Error

Behaviors/performance rated more alike or dissimilar than actually are b/c of info bias from another source

Duration Recording

Can be quantitative or qualitative length of time of occurrence Amount of time needed for completion length of time on or off task Latency (time behavior is not observed) Document time

Interview Examples

Canadian Occupational Performance Measure (COPM) Occupational Performance History Interview-II (OPHI-II)

Standardization occurs through

Clarification of procedures for administration and scoring Verification of psychometric properties Development of standard criteria or norms

Item Bias determined by

Classical test theory Item response theory (Rasch)

Occupational Performance Example

Cognitive Performance Test (CPT)

Rate Recording

Combo of event and duration recording Document frequency of occurrence by length of time

Qualitative Data Collection Methods for Observation

Detailed description of performance/behaviors observed

Improved client and program outcomes are caused by

Effective decision making in evaluation phase which leads to a clear intervention plan

Error: Halo Effect

Evaluators general impression of the test taker affects evaluators rating of performance

Error: Contrast Error

Evaluators subjective response to certain characteristic of the test taker influences scoring

Collection methods for observation: Types of Quantitative Methods

Event Recording Duration Recording Rate Recording Time Sampling

Analyzing Assessment

FIRST: evaluate validity SECOND: evaluate reliability

Test Sensitivity

How well does instrument identify all those who possess behavior or characteristic is question minimizes false negatives should be high

Test Specificity

Implies that all those manifesting target behavior or characteristic will be identified Minimizes false positives Vital in diagnostic assessments should be high

Types of Non-Standardized Assessments

Interview Questionnaire Observation Occupational Performance Assessment

Types of Errors in Standardized Assessments

Item Bias Participant Variable (Evaluator and Test taker)

What do we need to know about a specific problem

Its parameters/characteristics (what happens, when, how affects functioning, how much assistance)

Normative Assessments in OT

Mini Mental (MMSE) Sensory integration and praxis test (SIPT) Lowenstein OT cognitive assessment (LOTCA) Assessment of motor and process skills (AMPS)

Evaluator Effects on Testing

Observer presence (create inaccurate results) Observer Expectation/Observer Bias (create inaccurate results, severity/leniency, central tendency)

Questionnaire Examples

Occupational Self Assessment Version 2.2 (OSA) Child Occupational Self Assessment Version 2.1 (COSA)

Item Bias

Occurs when individuals of similar abilities perform differently on test item -Can be b/c of content asked or covered, language used

Error: Ambiguity Error

Raters interpretation of response to a test varies from typical interpretation made by others

Summative

Refers to the conclusions drawn from the evaluation or decisions made after a course of intervention (can occur during intervention)

Formative

Refers to the process of data gathering to identify the potential problem areas

What psychometric properties need to be evaluated for criterion referenced test

Reliability Validity Limitations (sample size, demographics)

What psychometric properties need to be evaluated for normative test

Reliability Validity Sample size and demographics Standard Error of Measurement

Event Recording

Simplest, most common Tally occurrences Document frequency of phenomenon

Types of Test Reliability

Test Re-test (stability) Interrater and intrarater Alternate, parallel Internal consistency

Pragmatic Reasoning

To consider practical constraints and social factors

Conditional Reasoning

To integrate all of the pieces specific to person

Interactive Reasoning

To respond to patients cues

Test Taker Variables

Tolerance for pain or discomfort Fatigue/stamina Energy Motivation Anxiety/stress Depression Paranoia Beliefs and views of testing

Ecological Validity

Two components: Verdicality: how well does it predict real-world performance Verisimilitude: how similar is test to real-world conditions

Alternate or Parallel Form

Two equivalent forms of the same test Good if items can be easily remembered or simulate learning

Construct Validity

Types: Convergent and discriminant -tests "goodness of fit" between theorized construct and resulting data from test

Criterion-Related Validity

Types: concurrent and predictive -compares test to external criterion

Norm-referenced tests can be used to establish

ability or disability in relation to normative sample justify services

Factorial Validity

ability to identify interrelated behaviors, abilities, or functions -factor analysis, multi-trait validity

Error: Proximity Error

an unexpected event immediately preceding an assessment triggers a reaction to test item

Clinical Responsiveness

can detect change when change happens

Reliability Measures:Below .70

caution, inadequate or unacceptable corrleation

Scorer Reliability: Intra-rater

consistency in measurement and scoring by evaluator when two test results from two similar situations are correlated

Scorer Reliability: Inter-rater

degree of agreement between two scores from two raters following observation and rating of same subject -(.85 or higher is favorable)

Discriminant Validity

degree to which two tests being conducted are measuring a DIFFERENT construct one group does better than the other Negative correlation looking at opposites (two different constructs)

Convergent Validity

degree to which two tests being conducted are measuring the SAME construct both groups do well Positive correlation

Scientific Reasoning

formulate provisional ideas (diagnosis, developmental milestones)

Evaluation needs to pinpoint and identify

functional skills and problems of individual, current environment and expected environment, persons cultural practices and personal meaning

Rasch Theory

gives continuum -easiest to hardest -should hug line

Validity Measures: .60-.80

high correlation

Reliability Measures: .90-.99

high correlation, preferred

Correlation co-efficient

indicates the degree of agreement/relationship between two measures

Validity Measures: .20- .40

low correlation

Validity Measures: .40-.60

moderate correlation

Time Sampling

more common in research recording behavior at specific intervals Document frequency related to time intervals

Validity Measures: .00-.20

negligible correlation

Internal Consistency

reflects the degree that test items measure the same construct (should be >.80) Types: split half, kuder richardson, coefficient alpha, rasch analysis "one of these things is not like the other"

Reliability Measures:.80-.89

satisfactory or adequate correlation

Narrative Reasoning

to construct a picture of individuals needs and situation (life story, qualitative)

Validity Measures: .80-1.00

very high correlation

Reliability Measures: .70-.79

weak or minimally acceptable correlation

Correlation

what is common between two measures


Ensembles d'études connexes

Affective disorders (connect quiz)

View Set

Chapter 45: Assessment and Management of Patients with Endocrine Disorders

View Set

Spinal Cord, Spinal Nerves, Reflexes

View Set

Prosocial Behavior (helping others)

View Set

18 Prepositions Extra Practice (Vince)

View Set

Section 4.2 Windows Basics Questions

View Set

Unit 6, Volcanic Catastrophes Wikipedia Homework

View Set