Assessment Centers Flashcards
• Correlations between ratings of different dimensions within the same exercise tend to have higher correlations (r = .58) than the correlations between ratings of the same dimension (or trait) across the various exercises (r = .25)• Competency demand hypothesis: when individuals are motivated to meet a situational demand and have competencies to do so, most will engage in the dominant response and there will be few individual differences• When situation demands responses that are more difficult (not everyone possesses the competency) differences will be more easily observed• Perhaps we need to get rid of the idea that something is wrong when traits measured in different situations (exercises) do not converge, as these exercises were designed to tap relatively distinct aspects of the job
Haaland & Christiansen (2002)
Assessor-related factors in assessment centers• Model 1- assessors possess limited information-processing capacities and therefore are not always able to meet the cognitive demands of the assessment center process• Model 2 (expert assessor model) - in which differences between novices and experts account for differences in rating quality• Assessee-related factors: cross-exercise consistency and dimensional (w/in exercise) differentiation• Evidence for convergent validity is established when assessors rate candidates who perform consistently across exercises. Evidence for discriminant validity is found when assessors rate candidates who perform differently on dimensions.• Careful AC design and assessor reliability are necessary but insufficient for establishing evidence of convergent and discriminant validity in AC. This is because candidate perf. may be limiting factors to establish evidence for construct validity.
Lievens (2002)
• Goal: look at criterion-validity of AC at dimension level• Past res. Collapse dimensions of AC to predict perf => deflated validity• Mean validity of 6 AC dimensions =.36• Problem-solving, influencing others, org. & planning, comm. =>best predictor of perf. ->drive & consideration/awareness X good predictors• Found above dimension intercorrelations =>need to increase discrimination among dimensions
Arthur et al (2003)
• Model of perf. Of 8 competencies• Leading & deciding; interacting & presenting; analyzing & interpreting; supporting & cooperating; creating & conceptualizing; adapting & coping; enterprising & performing; organizing & executing• Cog ability & personality (to a lesser extent) predicts great 8• Best predictor of perf =>analyzing & interpreting, org & executing; leading & deciding
Bartram (2005)
• Goal: provide factors influencing quality of assessor decisions• More than 4/5 dimensions will overburden daters• Assessee characteristics- whites rated higher in ex. w. cog. Component• Variance in ratings can be attribute to exercise format• Org. culture & context influence rating process too• Rating models: rational model, limited capacity model, expert model• Poor quality ratings can be due to 1)lack of opp to obs, 2) assessor biases, 3)assessor capability to make accurate judgments
Lievens & Klimoski (2001)
• Goal: assessor type & assessee perf profiles influencing construct validity of AC ratings• Assessor type =>limited capacity model & expert model• Assessee perf. Profiles =>1) consistent & diff. perf. 2) Consistent & undiff. Perf3) Inconsistent & undiff perf.4) Inconsistent & diff. perf• Increase convergent validity when rating assessee (1) & (2)• Increase discriminant validity when rating assessee (1)• Careful AC design & increase interrater reliability is not sufficient =>construct validity influenced by assessee perf. profiles
Lievens (2002)
book on ACs
Thornton & Rupp (2006)
• Exercise can be presented to diff participants at diff times – order that exercises are presented had trivial effects on the ratings
Bycio & Zoogah (2002)
• Despite various design fixes, “construct validity” as conceived in MTMM (multi-trait, multi-method) cannot be salvaged• Exercise effects are robust• Exercise effects represent cross-situational specificity and not method bias• Assessors can be quite accurate judges of behavior• Exercise effects do not indicate assessor halo error but represent accurate overall judgments of perf• ACs are good for exercise not dimension effects
Lance (2008)
• The evidence clearly confirms that the AC method does possess validity to assess and develop job-related perf dimensions• Dimensions should remain the focal construct• Exercise effects may be meaningful but dimensions can add to it
Rupp et al (2008)
• AC are not valid because are psychometrically unsound• Should use 1) test of traits 2) simulation exercises 3) bio indicators
Schuler (2008)
• Data from ACs vary in quality• ACs work when judgments made reflect useful insights that both the individual and the org find valuable…
Moses (2008)
• Lance (2008) ignores the multiple functions that the AC method fulfills and the reasons why ACs have been around for so long.
Jones & Klimoski (2008)
• Assessment Center Purposes: Employee selection, Identification of managerial talent, Development planning, Identification of training needs, Promotion, Management Succession• Selection centers: assessors serve more often, skills stay more current, assessors asked to create overall performance ratings, center data are validated more frequently• Development centers: fewer candidate selection mechanisms used, heavy reliance on supervisor data, many female assessees, varied exercises, assessors conduct long discussion sessions, center data are infrequently validated• Promotion centers: candidates selected from wider variety of sources, assessors asked to observe many more assesses, assessees complete fewer types of exercises
Spychalski et al. 1997
• Arthur et al – shouldn’t abandon dimension-based app because issue is failure to engage in appropriate tests of theory• Lack of appropriate test of espoused construct underlying dimension. Explains poor construct validity of AC• PEDR are item-level ratings and PCOR are composite scale level indicator of dimensions. Surprising PEDR reflect exercise effect and small dimension effect – typical of items in scale• Future directions: hold AC developers to same psychometric standards as test developers, design & implement AC that properly represent constructs.
Arthur et alResponse to Lance (2008)