😶‍🌫️
Psych
  • Preface
  • [4/9/2025] A One-Stop Calculator and Guide for 95 Effect-Size Variants
  • [4/9/2025] the people make the place
  • [4/9/2025] Personality predicts things
  • [3/31/2025] Response surface analysis with multilevel data
  • [3/11/2025] A Complete Guide to Natural Language Processing
  • [3/4/2025] Personality - Self and Identity
  • [3/1/2025] Updating Vocational Interests Information
  • [2/25/2025] Abilities & Skills
  • [2/22/2025] APA table format
  • [2/19/2025] LLM that replace human participants can harmfully misportray and flatt
  • [2/18/2025] Research Methods Knowledge Base
  • [2/17/2025] Personality - Motives/Interests
  • [2/11/2025] Trait structure
  • [2/10/2025] Higher-order construct
  • [2/4/2025] RL for CAT
  • [2/4/2025] DoWhy | An end-to-end library for causal inference
  • [2/4/2025] DAGitty — draw and analyze causal diagrams
  • [2/2/2025] Personality States
  • [2/2/2025] Psychometric Properties of Automated Video Interview Competency Assessments
  • [2/2/2025] How to diagnose abhorrent science
  • [1/28/2025] LLM and personality/interest items
  • [1/28/2025] Personality - Dispositions
  • [1/28/2025] Causal inference in statistics
  • [1/27/2025] Personality differences between birth order categories and across sibship sizes
  • [1/27/2025] nomological network meta-analysis.
  • [1/25/2025] Classic Papers on Scale Development/Validation
  • [1/17/2025] Personality Reading
  • [1/15/2025] Artificial Intelligence: Redefining the Future of Psychology
  • [1/13/2025] R for Psychometics
  • [12/24/2024] Comparison of interest congruence indices
  • [12/24/2024] Most recent article on interest fit measures
  • [12/24/2024] Grammatical Redundancy in Scales: Using the “ConGRe” Process to Create Better Measures
  • [12/24/2024] Confirmatory Factor Analysis with Word Embeddings
  • [12/24/2024] Can ChatGPT Develop a Psychometrically Sound Situational Judgment Test?
  • [12/24/2024] Using NLP to replace human content coders
  • [11/21/2024] AI Incident Database
  • [11/20/2024] Large Language Model-Enhanced Reinforcement Learning
  • [11/05/2024] Self-directed search
  • [11/04/2024] Interview coding and scoring
  • [11/04/2024] What if there were no personality factors?
  • [11/04/2024] BanditCAT and AutoIRT
  • [10/29/2024] LLM for Literature/Survey
  • [10/27/2024] Holland's Theory of Vocational Choice and Adjustment
  • [10/27/2024] Item Response Warehouse
  • [10/26/2024] EstCRM - the Samejima's Continuous IRT Model
  • [10/23/2024] Idiographic Personality Gaussian Process for Psychological Assessment
  • [10/23/2024] The experience sampling method (ESM)
  • [10/21/2024] Ecological Momentary Assessment (EMA)
  • [10/20/2024] Meta-Analytic Structural Equation Modeling
  • [10/20/2024] Structure of vocational interests
  • [10/17/2024] LLMs for psychological assessment
  • [10/16/2024] Can Deep Neural Networks Inform Theory?
  • [10/16/2024] Cognition & Decision Modeling Laboratory
  • [10/14/2024] Time-Invariant Confounders in Cross-Lagged Panel Models
  • [10/13/2024] Polynomial regression
  • [10/13/2024] Bayesian Mixture Modeling
  • [10/10/2024] Response surface analysis (RSA)
  • [10/10/2024] Text-Based Personality Assessment with LLM
  • [10/09/2024] Circular unidimensional scaling: A new look at group differences in interest structure.
  • [10/07/2024] Video Interview
  • [10/07/2024] Relationship between Measurement and ML
  • [10/07/2024] Conscientiousness × Interest Compensation (CONIC) model
  • [10/03/2024] Response modeling methodology
  • [10/02/2024] Conceptual Versus Empirical Distinctions Among Constructs
  • [10/02/2024] Construct Proliferation
  • [09/23/2024] Psychological Measurement Paradigm through Interactive Fiction Games
  • [09/20/2024] A Computational Method to Reveal Psychological Constructs From Text Data
  • [09/18/2024] H is for Human and How (Not) To Evaluate Qualitative Research in HCI
  • [09/17/2024] Automated Speech Recognition Bias in Personnel Selection
  • [09/16/2024] Congruency Effect
  • [09/11/2024] privacy, security, and trust perceptions
  • [09/10/2024] Measurement, Scale, Survey, Questionnaire
  • [09/09/2024] Reporting Systematic Reviews
  • [09/09/2024] Evolutionary Neuroscience
  • [09/09/2024] On Personality Measures and Their Data
  • [09/09/2024] Two Dimensions of Professor-Student Rapport Differentially Predict Student Success
  • [09/05/2024] The SAPA Personality Inventory
  • [09/05/2024] Moderated mediation
  • [09/03/2024] BiGGen Bench
  • [09/02/2024] LMSYS Chatbot Arena
  • [09/02/2024] Introduction to Measurement Theory Chapters 1, 2 (2.1-2.8) and 3.
  • [09/01/2024] HCI measurememt
  • [08/30/2024] Randomization Test
  • [08/30/2024] Interview Quantative Statistical
  • [08/29/2024] Cascading Model
  • [08/29/2024] Introduction: The White House (IS_202)
  • [08/29/2024] Circular unidimensional scaling
  • [08/28/2024] Sex and Gender Differences (Neur_542_Week2)
  • [08/26/2024] Workplace Assessment and Social Perceptions (WASP) Lab
  • [08/26/2024] Computational Organizational Research Lab
  • [08/26/2024] Reading List (Recommended by Bo)
  • [08/20/2024] Illinois NeuroBehavioral Assessment Laboratory (INBAL)
  • [08/14/2024] Quantitative text analysis
  • [08/14/2024] Measuring complex psychological and sociological constructs in large-scale text
  • [08/14/2024] LLM for Social Science Research
  • [08/14/2024] GPT for multilingual psychological text analysis
  • [08/12/2024] Questionable Measurement Practices and How to Avoid Them
  • [08/12/2024] NLP for Interest (from Dan Putka)
  • [08/12/2024] ONet Interest Profiler (Long and Short Scale)
  • [08/12/2024] ONet Interests Data
  • [08/12/2024] The O*NET-SOC Taxonomy
  • [08/12/2024] ML Ratings for O*Net
  • [08/09/2024] Limited ability of LLMs to simulate human psychological behaviours
  • [08/08/2024] A large-scale, gamified online assessment
  • [08/08/2024] Text-Based Traitand Cue Judgments
  • [08/07/2024] Chuan-Peng Lab
  • [08/07/2024] Modern psychometrics: The science of psychological assessment
  • [08/07/2024] Interactive Survey
  • [08/06/2024] Experimental History
  • [08/06/2024] O*NET Research reports
  • [07/30/2024] Creating a psychological assessment tool based on interactive storytelling
  • [07/24/2024] My Life with a Theory
  • [07/24/2024] NLP for Interest Job Ratings
  • [07/17/2024] Making vocational choices
  • [07/17/2024] Taxonomy of Psychological Situation
  • [07/12/2024] PathChat 2
  • [07/11/2024] Using games to understand the mind
  • [07/10/2024] Gamified Assessments
  • [07/09/2024] Poldracklab Software and Data
  • [07/09/2024] Consensus-based Recommendations for Machine-learning-based Science
  • [07/08/2024] Using AI to assess personal qualities
  • [07/08/2024] AI Psychometrics And Psychometrics Benchmark
  • [07/02/2024] Prompt Engineering Guide
  • [06/28/2024] Observational Methods and Qualitative Data Analysis 5-6
  • [06/28/2024] Observational Methods and Qualitative Data Analysis 3-4
  • [06/28/2024] Interviewing Methods 5-6
  • [06/28/2024] Interviewing Methods 3-4
  • [06/28/2024] What is Qualitative Research 3
  • [06/27/2024] APA Style
  • [06/27/2024] Statistics in Psychological Research 6
  • [06/27/2024] Statistics in Psychological Research 5
  • [06/23/2024] Bayesian Belief Network
  • [06/18/2024] Fair Comparisons in Heterogenous Systems Evaluation
  • [06/18/2024] What should we evaluate when we use technology in education?
  • [06/16/2024] Circumplex Model
  • [06/12/2024] Ways of Knowing in HCI
  • [06/09/2024] Statistics in Psychological Research 1-4
  • [06/08/2024] Mathematics for Machine Learning
  • [06/08/2024] Vocational Interests SETPOINT Dimensions
  • [06/07/2024] How's My PI Study
  • [06/06/2024] Best Practices in Supervised Machine Learning
  • [06/06/2024] SIOP
  • [06/06/2024] Measurement, Design, and Analysis: An Integrated Approach (Chu Recommended)
  • [06/06/2024] Classical Test Theory
  • [06/06/2024] Introduction to Measurement Theory (Bo Recommended)
  • [06/03/2024] EDSL: AI-Powered Research
  • [06/03/2024] Perceived Empathy of Technology Scale (PETS)
  • [06/02/2024] HCI area - Quantitative and Qualitative Modeling and Evaluation
  • [05/26/2024] Psychometrics with R
  • [05/26/2024] Programming Grammer Design
  • [05/25/2024] Psychometric Network Analysis
  • [05/23/2024] Item Response Theory
  • [05/22/2024] Nature Human Behaviour (Jan - 20 May, 2024)
  • [05/22/2024] Nature Human Behaviour - Navigating the AI Frontier
  • [05/22/2024] Computer Adaptive Testing
  • [05/22/2024] Personality Scale (Jim Shard)
  • [05/22/2024] Reliability
  • [05/19/2024] Chatbot (Jim Shared)
  • [05/17/2024] GOMS and Keystroke-Level Model
  • [05/17/2024] The Psychology of Human-Computer Interaction
  • [05/14/2024] Computational Narrative (Mark's Group)
  • [05/14/2024] Validity Coding
  • [05/14/2024] LLM as A Evaluator
  • [05/14/2024] Social Skill Training via LLMs (Diyi's Group)
  • [05/14/2024] AI Persona
  • [05/09/2024] Psychological Methods Journal Sample Articles
  • [05/08/2024] Meta-Analysis
  • [05/07/2024] Mturk
  • [05/06/2024] O*NET Reports and Documents
  • [05/04/2024] NLP and Chatbot on Personality Assessment (Tianjun)
  • [05/02/2024] Reads on Construct Validation
  • [04/25/2024] Reads on Validity
  • [04/18/2024] AI for Assessment
  • [04/17/2024] Interest Assessment
  • [04/16/2024] Personality Long Reading List (Jim)
    • Personality Psychology Overview
      • Why Study Personality Assessment
    • Dimensions and Types
    • Reliability
    • Traits: Two Views
    • Validity--Classical Articles and Reflections
    • Validity-Recent Proposals
    • Multimethod Perspective and Social Desirability
    • Paradigm of Personality Assessment: Multivariate
    • Heritability of personality traits
    • Classical Test-Construction
    • IRT
    • Social desirability in scale construction
    • Traits and culture
    • Paradigms of personality assessment: Empirical
    • Comparison of personality test construction strategies
    • Clinical versus Actuarial (AI) Judgement and Diagnostics
    • Decisions: Importance of base rates
    • Paradigms of Personality Assessment: Psychodynamic
    • Paradigms of Assessment: Interpersonal
    • Paradigms of Personality Assessment: Personological
    • Retrospective reports
    • Research Paradigms
    • Personality Continuity and Change
Powered by GitBook
On this page

[05/02/2024] Reads on Construct Validation

Previous[05/04/2024] NLP and Chatbot on Personality Assessment (Tianjun)Next[04/25/2024] Reads on Validity

Last updated 1 year ago

Flake, J. K., Pek, J., & Hehman, E. (2017). Construct validation in social and personality research: Current practice and recommendations. Social Psychological and Personality Science

Abstract

The verity of results about a psychological construct hinges on the validity of its measurement, making construct validation a fundamental methodology to the scientific process. We reviewed a representative sample of articles published in the Journal of Personality and Social Psychology for construct validity evidence. We report that latent variable measurement, in which responses to items are used to represent a construct, is pervasive in social and personality research. However, the field does not appear to be engaged in best practices for ongoing construct validation. We found that validity evidence of existing and author-developed scales was lacking, with coefficient α often being the only psychometric evidence reported. We provide a discussion of why the construct validation framework is important for social and personality researchers and recommendations for improving practice.

Summary

The authors reviewed articles from the Journal of Personality and Social Psychology to assess the current practices of construct validation. They found that while latent variable measurement is common in the field, best practices for ongoing construct validation are lacking. The document emphasizes the significance of construct validation in ensuring the validity of measurement in psychological research and provides recommendations for improving practices in this area. Key points and arguments from the document include:

  • Construct validation is crucial for ensuring the validity of measurement in psychological research.

  • The field of social and personality research often relies on latent variable measurement to represent constructs.

  • The document highlights the lack of validity evidence for existing and author-developed scales, with coefficient alpha often being the only reported psychometric evidence.

  • Recommendations are provided for improving construct validation practices in social and personality research.

  • The document emphasizes the importance of ongoing construct validation to support the verity of results in psychological research.

Key Findings:

  • Latent variable measurement is pervasive in social and personality research.

  • Lack of validity evidence for existing and author-developed scales.

  • Recommendations for improving construct validation practices.

  • Importance of ongoing construct validation for supporting research findings.

Takeaways

How prevalent is latent variable measurement in social research?

  • Latent variable measurement is pervasive in social and personality research, with responses to items commonly used to represent constructs in the field. On average, latent variable measurement accounted for 87% of the measures used in social and personality psychology research.

What are common constructs represented by latent variables in research?

  • Common constructs represented by latent variables in social and personality research include attitudes, life satisfaction, status, and various psychological phenomena that are typically unobservable (such as attitudes).

What are examples of latent variables in psychological research?

  • In psychological research, examples of latent variables include attitudes, life satisfaction, status, and various psychological phenomena that are typically unobservable. These latent constructs are often measured using scales, surveys, questionnaires, and tests, where responses to items are used to represent the underlying construct. The process of construct validation is crucial in ensuring that the measures appropriately represent the latent variables of interest

Messick, S. (1995). Validity of psychological assessment: Validation of inferences from persons’ responses and performances as scientific inquiry into score meaning. American Psychologist, 50, 741-749.

Summary

The author discusses the concept of validity in psychological assessment, emphasizing the need for a unified approach that considers not only traditional validity types (content, criterion, and construct) but also the value implications of score meaning and the social consequences of score use. The new unified concept of validity integrates these aspects into a construct framework for testing hypotheses about score meaning and relationships. Six aspects of construct validity are highlighted: content, substantive, structural, generalizability, external, and consequential. The document stresses the importance of validity in all assessments, including performance assessments, and addresses sources of invalidity such as construct underrepresentation and construct-irrelevant variance. It also discusses the importance of evidence in construct validity and the need to evaluate the fit of information with theoretical rationales for score interpretation.

  • Traditional validity types: content, criterion, and construct

  • Unified concept of validity integrating value implications and social consequences

  • Six aspects of construct validity: content, substantive, structural, generalizability, external, and consequential

  • Importance of validity in all assessments, including performance assessments

  • Sources of invalidity: construct underrepresentation and construct-irrelevant variance

  • Importance of evidence in construct validity and evaluating fit with theoretical rationales

Takeaways

How does the unified concept of validity enhance assessment practices?

  • The unified concept of validity enhances assessment practices by providing a comprehensive framework that integrates various aspects of validity to address the complexities inherent in appraising the appropriateness, meaningfulness, and usefulness of score inferences. This unified concept highlights six distinguishable aspects of construct validity: content, substantive, structural, generalizability, external, and consequential, which serve as general validity criteria for all educational and psychological measurement. By considering these aspects, researchers and practitioners can ensure that the evidence and inferences drawn from test scores are well-supported and aligned with theoretical rationales, leading to more robust score interpretations and actions. Additionally, the unified concept of validity emphasizes the importance of integrating evidence related to both score meaning and value implications, thereby promoting a more holistic approach to test validation that considers the social consequences and ethical implications of assessment practices

What are the six aspects of construct validity?

  • Content aspect, which includes evidence of content relevance, representativeness, and technical quality

  • Substantive aspect, which refers to theoretical rationales for observed consistencies in test responses and empirical evidence that theoretical processes are engaged by respondents

  • Structural aspect, which appraises the fidelity of scoring structure to the structure of the construct domain at issue

  • Generalizability aspect, which examines the extent to which score properties and interpretations generalize to and across population groups, settings, and tasks

  • External aspect, which includes convergent and discriminant evidence from multitrait-multimethod comparisons and evidence of criterion relevance and applied utility

  • Consequential aspect, which appraises the value implications of score interpretation as a basis for action and the actual and potential consequences of test use

How does the consequential aspect relate to construct validity?

  • The consequential aspect of construct validity is closely related to the overall construct validity framework. This aspect involves evaluating the value implications of score interpretation as a basis for action and assessing the actual and potential consequences of test use, especially in terms of sources of invalidity related to issues of bias, fairness, and distributive justice. The consequential aspect considers both the intended and unintended consequences of score interpretation and use, including short- and long-term effects. It examines the social consequences of testing, which can be either positive or negative, such as improved educational policies based on international comparisons of student performance or negative impacts associated with bias in scoring and interpretation. The consequential aspect is essential in ensuring that the outcomes of test interpretation and use align with the intended purposes and do not lead to adverse effects resulting from sources of test invalidity.

What are the social consequences of testing in educational settings?

  • In educational settings, the social consequences of testing can have both positive and negative impacts. Positive consequences may include benefits such as improved teaching and learning practices, while negative consequences could arise from sources of test invalidity, such as construct underrepresentation or construct-irrelevant variance. It is crucial to assess both potential and actual social consequences of testing to ensure that any adverse effects are not attributable to test invalidity and to make informed decisions about test interpretation and use.

203KB
flake-et-al-2017-construct-validation-in-social-and-personality-research-current-practice-and-recommendations.pdf
pdf
2MB
validity of psychological assessment.pdf
pdf