😶‍🌫️
Psych
  • Preface
  • [4/9/2025] A One-Stop Calculator and Guide for 95 Effect-Size Variants
  • [4/9/2025] the people make the place
  • [4/9/2025] Personality predicts things
  • [3/31/2025] Response surface analysis with multilevel data
  • [3/11/2025] A Complete Guide to Natural Language Processing
  • [3/4/2025] Personality - Self and Identity
  • [3/1/2025] Updating Vocational Interests Information
  • [2/25/2025] Abilities & Skills
  • [2/22/2025] APA table format
  • [2/19/2025] LLM that replace human participants can harmfully misportray and flatt
  • [2/18/2025] Research Methods Knowledge Base
  • [2/17/2025] Personality - Motives/Interests
  • [2/11/2025] Trait structure
  • [2/10/2025] Higher-order construct
  • [2/4/2025] RL for CAT
  • [2/4/2025] DoWhy | An end-to-end library for causal inference
  • [2/4/2025] DAGitty — draw and analyze causal diagrams
  • [2/2/2025] Personality States
  • [2/2/2025] Psychometric Properties of Automated Video Interview Competency Assessments
  • [2/2/2025] How to diagnose abhorrent science
  • [1/28/2025] LLM and personality/interest items
  • [1/28/2025] Personality - Dispositions
  • [1/28/2025] Causal inference in statistics
  • [1/27/2025] Personality differences between birth order categories and across sibship sizes
  • [1/27/2025] nomological network meta-analysis.
  • [1/25/2025] Classic Papers on Scale Development/Validation
  • [1/17/2025] Personality Reading
  • [1/15/2025] Artificial Intelligence: Redefining the Future of Psychology
  • [1/13/2025] R for Psychometics
  • [12/24/2024] Comparison of interest congruence indices
  • [12/24/2024] Most recent article on interest fit measures
  • [12/24/2024] Grammatical Redundancy in Scales: Using the “ConGRe” Process to Create Better Measures
  • [12/24/2024] Confirmatory Factor Analysis with Word Embeddings
  • [12/24/2024] Can ChatGPT Develop a Psychometrically Sound Situational Judgment Test?
  • [12/24/2024] Using NLP to replace human content coders
  • [11/21/2024] AI Incident Database
  • [11/20/2024] Large Language Model-Enhanced Reinforcement Learning
  • [11/05/2024] Self-directed search
  • [11/04/2024] Interview coding and scoring
  • [11/04/2024] What if there were no personality factors?
  • [11/04/2024] BanditCAT and AutoIRT
  • [10/29/2024] LLM for Literature/Survey
  • [10/27/2024] Holland's Theory of Vocational Choice and Adjustment
  • [10/27/2024] Item Response Warehouse
  • [10/26/2024] EstCRM - the Samejima's Continuous IRT Model
  • [10/23/2024] Idiographic Personality Gaussian Process for Psychological Assessment
  • [10/23/2024] The experience sampling method (ESM)
  • [10/21/2024] Ecological Momentary Assessment (EMA)
  • [10/20/2024] Meta-Analytic Structural Equation Modeling
  • [10/20/2024] Structure of vocational interests
  • [10/17/2024] LLMs for psychological assessment
  • [10/16/2024] Can Deep Neural Networks Inform Theory?
  • [10/16/2024] Cognition & Decision Modeling Laboratory
  • [10/14/2024] Time-Invariant Confounders in Cross-Lagged Panel Models
  • [10/13/2024] Polynomial regression
  • [10/13/2024] Bayesian Mixture Modeling
  • [10/10/2024] Response surface analysis (RSA)
  • [10/10/2024] Text-Based Personality Assessment with LLM
  • [10/09/2024] Circular unidimensional scaling: A new look at group differences in interest structure.
  • [10/07/2024] Video Interview
  • [10/07/2024] Relationship between Measurement and ML
  • [10/07/2024] Conscientiousness × Interest Compensation (CONIC) model
  • [10/03/2024] Response modeling methodology
  • [10/02/2024] Conceptual Versus Empirical Distinctions Among Constructs
  • [10/02/2024] Construct Proliferation
  • [09/23/2024] Psychological Measurement Paradigm through Interactive Fiction Games
  • [09/20/2024] A Computational Method to Reveal Psychological Constructs From Text Data
  • [09/18/2024] H is for Human and How (Not) To Evaluate Qualitative Research in HCI
  • [09/17/2024] Automated Speech Recognition Bias in Personnel Selection
  • [09/16/2024] Congruency Effect
  • [09/11/2024] privacy, security, and trust perceptions
  • [09/10/2024] Measurement, Scale, Survey, Questionnaire
  • [09/09/2024] Reporting Systematic Reviews
  • [09/09/2024] Evolutionary Neuroscience
  • [09/09/2024] On Personality Measures and Their Data
  • [09/09/2024] Two Dimensions of Professor-Student Rapport Differentially Predict Student Success
  • [09/05/2024] The SAPA Personality Inventory
  • [09/05/2024] Moderated mediation
  • [09/03/2024] BiGGen Bench
  • [09/02/2024] LMSYS Chatbot Arena
  • [09/02/2024] Introduction to Measurement Theory Chapters 1, 2 (2.1-2.8) and 3.
  • [09/01/2024] HCI measurememt
  • [08/30/2024] Randomization Test
  • [08/30/2024] Interview Quantative Statistical
  • [08/29/2024] Cascading Model
  • [08/29/2024] Introduction: The White House (IS_202)
  • [08/29/2024] Circular unidimensional scaling
  • [08/28/2024] Sex and Gender Differences (Neur_542_Week2)
  • [08/26/2024] Workplace Assessment and Social Perceptions (WASP) Lab
  • [08/26/2024] Computational Organizational Research Lab
  • [08/26/2024] Reading List (Recommended by Bo)
  • [08/20/2024] Illinois NeuroBehavioral Assessment Laboratory (INBAL)
  • [08/14/2024] Quantitative text analysis
  • [08/14/2024] Measuring complex psychological and sociological constructs in large-scale text
  • [08/14/2024] LLM for Social Science Research
  • [08/14/2024] GPT for multilingual psychological text analysis
  • [08/12/2024] Questionable Measurement Practices and How to Avoid Them
  • [08/12/2024] NLP for Interest (from Dan Putka)
  • [08/12/2024] ONet Interest Profiler (Long and Short Scale)
  • [08/12/2024] ONet Interests Data
  • [08/12/2024] The O*NET-SOC Taxonomy
  • [08/12/2024] ML Ratings for O*Net
  • [08/09/2024] Limited ability of LLMs to simulate human psychological behaviours
  • [08/08/2024] A large-scale, gamified online assessment
  • [08/08/2024] Text-Based Traitand Cue Judgments
  • [08/07/2024] Chuan-Peng Lab
  • [08/07/2024] Modern psychometrics: The science of psychological assessment
  • [08/07/2024] Interactive Survey
  • [08/06/2024] Experimental History
  • [08/06/2024] O*NET Research reports
  • [07/30/2024] Creating a psychological assessment tool based on interactive storytelling
  • [07/24/2024] My Life with a Theory
  • [07/24/2024] NLP for Interest Job Ratings
  • [07/17/2024] Making vocational choices
  • [07/17/2024] Taxonomy of Psychological Situation
  • [07/12/2024] PathChat 2
  • [07/11/2024] Using games to understand the mind
  • [07/10/2024] Gamified Assessments
  • [07/09/2024] Poldracklab Software and Data
  • [07/09/2024] Consensus-based Recommendations for Machine-learning-based Science
  • [07/08/2024] Using AI to assess personal qualities
  • [07/08/2024] AI Psychometrics And Psychometrics Benchmark
  • [07/02/2024] Prompt Engineering Guide
  • [06/28/2024] Observational Methods and Qualitative Data Analysis 5-6
  • [06/28/2024] Observational Methods and Qualitative Data Analysis 3-4
  • [06/28/2024] Interviewing Methods 5-6
  • [06/28/2024] Interviewing Methods 3-4
  • [06/28/2024] What is Qualitative Research 3
  • [06/27/2024] APA Style
  • [06/27/2024] Statistics in Psychological Research 6
  • [06/27/2024] Statistics in Psychological Research 5
  • [06/23/2024] Bayesian Belief Network
  • [06/18/2024] Fair Comparisons in Heterogenous Systems Evaluation
  • [06/18/2024] What should we evaluate when we use technology in education?
  • [06/16/2024] Circumplex Model
  • [06/12/2024] Ways of Knowing in HCI
  • [06/09/2024] Statistics in Psychological Research 1-4
  • [06/08/2024] Mathematics for Machine Learning
  • [06/08/2024] Vocational Interests SETPOINT Dimensions
  • [06/07/2024] How's My PI Study
  • [06/06/2024] Best Practices in Supervised Machine Learning
  • [06/06/2024] SIOP
  • [06/06/2024] Measurement, Design, and Analysis: An Integrated Approach (Chu Recommended)
  • [06/06/2024] Classical Test Theory
  • [06/06/2024] Introduction to Measurement Theory (Bo Recommended)
  • [06/03/2024] EDSL: AI-Powered Research
  • [06/03/2024] Perceived Empathy of Technology Scale (PETS)
  • [06/02/2024] HCI area - Quantitative and Qualitative Modeling and Evaluation
  • [05/26/2024] Psychometrics with R
  • [05/26/2024] Programming Grammer Design
  • [05/25/2024] Psychometric Network Analysis
  • [05/23/2024] Item Response Theory
  • [05/22/2024] Nature Human Behaviour (Jan - 20 May, 2024)
  • [05/22/2024] Nature Human Behaviour - Navigating the AI Frontier
  • [05/22/2024] Computer Adaptive Testing
  • [05/22/2024] Personality Scale (Jim Shard)
  • [05/22/2024] Reliability
  • [05/19/2024] Chatbot (Jim Shared)
  • [05/17/2024] GOMS and Keystroke-Level Model
  • [05/17/2024] The Psychology of Human-Computer Interaction
  • [05/14/2024] Computational Narrative (Mark's Group)
  • [05/14/2024] Validity Coding
  • [05/14/2024] LLM as A Evaluator
  • [05/14/2024] Social Skill Training via LLMs (Diyi's Group)
  • [05/14/2024] AI Persona
  • [05/09/2024] Psychological Methods Journal Sample Articles
  • [05/08/2024] Meta-Analysis
  • [05/07/2024] Mturk
  • [05/06/2024] O*NET Reports and Documents
  • [05/04/2024] NLP and Chatbot on Personality Assessment (Tianjun)
  • [05/02/2024] Reads on Construct Validation
  • [04/25/2024] Reads on Validity
  • [04/18/2024] AI for Assessment
  • [04/17/2024] Interest Assessment
  • [04/16/2024] Personality Long Reading List (Jim)
    • Personality Psychology Overview
      • Why Study Personality Assessment
    • Dimensions and Types
    • Reliability
    • Traits: Two Views
    • Validity--Classical Articles and Reflections
    • Validity-Recent Proposals
    • Multimethod Perspective and Social Desirability
    • Paradigm of Personality Assessment: Multivariate
    • Heritability of personality traits
    • Classical Test-Construction
    • IRT
    • Social desirability in scale construction
    • Traits and culture
    • Paradigms of personality assessment: Empirical
    • Comparison of personality test construction strategies
    • Clinical versus Actuarial (AI) Judgement and Diagnostics
    • Decisions: Importance of base rates
    • Paradigms of Personality Assessment: Psychodynamic
    • Paradigms of Assessment: Interpersonal
    • Paradigms of Personality Assessment: Personological
    • Retrospective reports
    • Research Paradigms
    • Personality Continuity and Change
Powered by GitBook
On this page

[04/25/2024] Reads on Validity

Cronbach, L. J. & Meehl, P. E. (1955). Construct validity in psychological tests. Psychological Bulletin, 52, 281-302.

Summary

This paper discusses the concept of construct validity in psychological tests, highlighting the importance of distinguishing between different types of validity in test validation studies.

  • Validation of psychological tests has not been adequately conceptualized, leading to the need for distinguishing four types of validity: predictive validity, concurrent validity, content validity, and construct validity.

  • Construct validity is crucial when interpreting a test as a measure of an attribute that is not operationally defined, requiring the identification of psychological constructs that account for variance in test performance.

  • the importance of construct validation for various types of psychological tests, such as aptitude, achievement, and interests, and provides examples of how construct validity is involved in interpreting test results.

  • Constructs in test interpretation are typically postulated attributes of people reflected in test performance, and the logic of construct validation is invoked to defend proposed interpretations of tests.

  • discusses the critical view of the criterion implied in construct validity, highlighting the importance of examining the entire body of evidence offered when making claims about constructs measured by a test.

Takeaway

How to distinguish between different types of validity?

  • we can distinguish between different types of validity by focusing on the criterion used in the validation process.

  • We can distinguish between predictive and concurrent validity based on the timing of when the test scores and criterion scores are determined. Predictive validity involves studying the relationship between test scores and a criterion that is obtained after the test is given, while concurrent validity examines the correlation between test scores and criterion scores that are determined at the same time. In predictive validity, the criterion is obtained after the test, whereas in concurrent validity, both the test score and criterion score are determined simultaneously

  • Content validity is established by ensuring that test items represent a sample of the universe of interest, while construct validity is crucial when there is no clear criterion measure available, requiring the identification of underlying traits or qualities.

  • Construct validation is important for various psychological tests, such as aptitude, achievement, and interests, as it helps researchers understand what psychological constructs account for test performance.

How is predictive validity determined in relation to test scores?

  • Predictive validity in relation to test scores is determined by studying the relationship between the test scores and a criterion that is obtained after the test is given. This involves assessing how well the test scores predict future performance or outcomes. Predictive validity is particularly important in fields such as education and employment, where the ability of a test to predict future success is crucial.

How is the relationship between test scores and criteria studied?

  • The relationship between test scores and criteria is studied through various types of validity, such as predictive validity, content validity, and construct validity. Predictive validity involves assessing how well test scores predict future performance or outcomes. Content validity is studied when the tester is concerned with the type of behavior involved in the test performance. Construct validity is important for psychological tests and focuses on the underlying trait or quality being measured by the test. Studies of process, such as observing the person's performance during the test, can also provide insights into the relationship between test scores and criteria.

Why is construct validity important in psychological test relationships?

  • Construct validity is important in psychological test relationships because it focuses on the underlying trait or quality being measured by the test, rather than just the test behavior or scores on the criteria. It helps in answering questions such as whether a test is culture-free, what specific trait a test measures, and how individuals with different scores on the test differ from each other. Construct validation is crucial for various types of psychological tests, including aptitude, achievement, and interests tests, as it allows researchers to understand the theoretical basis of the test and its relationship to the construct being measured.

How does construct validity help differentiate individuals based on scores?

  • Construct validity helps differentiate individuals based on scores by focusing on the underlying trait or quality being measured by the test, rather than just the test behavior or scores on the criteria. This allows researchers to understand the theoretical basis of the test and its relationship to the construct being measured, which in turn helps in distinguishing individuals with different scores on the test and how they differ from each other

What is the role of construct validity in differentiating individuals?

  • Construct validity plays a crucial role in differentiating individuals based on scores by focusing on the underlying trait or quality being measured by the test, rather than just the test behavior or scores on the criteria. It helps researchers understand the theoretical basis of the test and its relationship to the construct being measured, allowing for a more nuanced differentiation among individuals with varying scores on the test. By examining the trait or quality underlying the test, construct validity enables researchers to make distinctions between individuals based on the specific construct being assessed, leading to more accurate and meaningful interpretations of test scores.

Can construct validity enhance the accuracy of test score interpretations?

  • Construct validity can indeed enhance the accuracy of test score interpretations by focusing on the underlying trait or quality being measured by the test, rather than just the test behavior or scores on the criteria. By examining the trait or quality underlying the test, construct validity enables researchers to make distinctions between individuals based on the specific construct being assessed, leading to more accurate and meaningful interpretations of test scores.

What is the focus of construct validity in test assessments?

  • The focus of construct validity in test assessments is on the underlying trait or quality being measured by the test, rather than just the test behavior or scores on the criteria. By examining the trait or quality underlying the test, construct validity enables researchers to make distinctions between individuals based on the specific construct being assessed, leading to more accurate and meaningful interpretations of test scores.

Why is understanding the underlying trait important for test assessments?

  • Understanding the underlying trait is important for test assessments because it allows researchers to identify the real dimensions of human personality and create pure tests that can predict complex behaviors accurately. Additionally, by examining the underlying trait, researchers can make distinctions between individuals based on the specific construct being assessed, leading to more accurate and meaningful interpretations of test scores.

Campbell, D. T. & Fiske, D. W. (1959). Convergent and discriminant validation by the multitrait-multimethod matrix. Psychological Bulletin, 56, 81-105.

Summary

This paper discusses the validation of measures of individual differences through a multitrait-multimethod matrix.

  • Validation is typically convergent, involving confirmation by independent measurement procedures.

  • Both convergent and discriminant validation are necessary for justifying novel trait measures, validating test interpretation, and establishing construct validity.

  • Each test or task used for measurement is a trait-method unit, combining trait content with measurement procedures.

  • To examine discriminant validity, multiple traits and methods must be employed, often through a multitrait-multimethod matrix.

  • The document provides a synthetic example of a multitrait-multimethod matrix to illustrate the validation process.

  • The importance of independence in validation processes is highlighted, emphasizing the distinction between reliability and validity.

  • Discriminant validation is crucial to distinguish between different constructs and ensure the validity of measurement instruments.

  • The concept of test as a trait-method unit is discussed, emphasizing the impact of method variance on test scores and the need to account for it in validation processes.

Takeaway

What is discriminant validation?

  • Disciminant validation is a process used to establish construct validity by demonstrating that a measure of a trait has little overlap with measures of other traits. It is essential for justifying novel trait measures, validating test interpretation, and establishing construct validity. Discriminant validation is required in addition to convergent validation to ensure that tests are not invalidated by high correlations with other tests intended to measure different traits.

What is the purpose of discriminant validation in research?

  • Disciminant validation in research serves the purpose of establishing construct validity by demonstrating that a measure of a trait has little overlap with measures of other traits. It is essential for justifying novel trait measures, validating test interpretation, and ensuring that tests are not invalidated by high correlations with other tests intended to measure different traits.

What are the implications of failing discriminant validation in research?

  • Failing discriminant validation in research can have significant implications. It can lead to the invalidation of tests due to high correlations with other tests purporting to measure different traits, as seen in the classic case of social intelligence tests. When values in the heterotrait-heteromethod triangles are as high as those in the validity diagonal, or when heterotrait values are as high as the reliabilities within a monomethod block, it can indicate invalidity. Additionally, failing discriminant validation may suggest that the measures of a trait overlap with measures of other traits, undermining the distinctiveness and validity of the construct being studied.

Why is both convergent and discriminant validation necessary?

  • Both convergent and discriminant validation are necessary for justifying novel trait measures, validating test interpretation, or establishing construct validity. Tests can be invalidated by high correlations with other tests they were intended to differ from, highlighting the importance of discriminant validation. Discriminant validation is crucial to distinguish between different constructs and ensure the validity of measurement instruments.

How does discriminant validation help establish construct validity?

  • Disciminant validation helps establish construct validity by verifying the distinctions between the new dimension or construct being proposed and other existing constructs. It is crucial to demonstrate that the measure of a trait has little overlap with measures of other traits, thus supporting the differentiation of the trait being measured

How can high correlations between tests affect test validity?

  • High correlations between tests can affect test validity by potentially leading to the invalidation of tests. When values in the heterotrait-heteromethod triangles are as high as those in the validity diagonal, or when heterotrait values are as high as the reliabilities within a monomethod block, it can indicate invalidity. Additionally, tests can be invalidated if they have too high correlations with other tests purporting to measure different traits, as seen in the classic case of social intelligence tests.

What is the impact of heterotrait-heteromethod triangles on test validity?

  • The impact of heterotrait-heteromethod triangles on test validity is significant as they play a crucial role in assessing discriminant validity. These triangles help in evaluating whether a variable correlates higher with an independent effort to measure the same trait than with measures designed to capture different traits using the same method. The presence of heterotrait-heteromethod triangles allows researchers to compare validity values with correlations in these triangles to ensure that variables align with the expected trait relationships, providing evidence for discriminant validity. Additionally, the heterotrait-heteromethod triangles help in demonstrating the independence and convergence of different methods used in the validation process, which is essential for establishing the validity of the test.

Fiske, D. W., & Campbell, D. T. (1992). Citations do not solve problems. Psychological Bulletin, 112, 393-395.

Summary

The paper discusses the ongoing challenges in the field of psychology related to the validation of measures and methods used in research.

  • Campbell and Fiske's 1959 article emphasized the importance of measures of the same variable agreeing better than measures of different variables made by different methods

  • Current multitrait-multimethod matrices show little improvement over examples from 1959

  • Conceptual and methodological issues regarding the linkage between variables and their measurements are raised

  • Lack of consensus on appropriate statistical analysis of these matrices

  • High citation rates of the original article, but unresolved problems persist

  • Need for further exploration and refinement in understanding the relationship between psychological methods and constructs

Loevinger, J. (1957). Objective tests as instruments of psychological theory. Psychological Reports, Monograph Supplement, 3, 635–694.

Summary

This paper discusses the extension of the concept of validity in psychometrics, particularly focusing on construct validity.

  • Introduction to the critique of classical validity concept and the elucidation of construct validity terms.

  • Discussion on the relationship between test behavior and theory, including test responses as signs and samples, homogeneity issues, and observation prior to measurement.

  • Exploration of the components of construct validity, such as substantive, structural, and external components.

  • Analysis of secular trends in test behavior and their implications for construct validity.

  • Consideration of alternative approaches to objective tests, including different kinds of data and analysis methods.

  • Overview of psychometrics theory, methods, and applications.

Takeaway

How does the paper redefine construct validity in psychometrics?

  • The paper redefines construct validity in psychometrics by emphasizing the importance of measuring traits that have real existence, rather than focusing solely on prediction or utility-oriented approaches. It argues that construct validity is the whole subject from a systematic, scientific point of view, challenging the traditional classification of validity and advocating for a distinction between administrative and scientific validity. The document also discusses the criteria for construct validity, highlighting the need for consistency in the substance, structure, and external correlations of the test to establish construct validity. Additionally, it stresses the importance of convergence of multiple lines of evidence to establish construct validity, including psychometric and psychological evidence.

  • The new emphasis in redefining construct validity is on measuring traits that have real existence, rather than solely focusing on prediction or utility-oriented approaches

What is the significance of measuring traits with real existence?

  • The significance of measuring traits with real existence lies in the idea that focusing on traits that have real existence can lead to a more fruitful development of psychometric devices and theory. This orientation is seen as antithetical to approaches that prioritize prediction, decisions, or utility. By measuring traits that have real existence, researchers can gain a deeper understanding of psychological phenomena and improve the validity of their assessments.

How does prioritizing real traits improve validity of assessments?

  • Prioritizing real traits in assessments can improve validity by ensuring that the measurements are grounded in actual psychological phenomena that exist independently of the measurement process itself. This approach helps in developing a deeper understanding of psychological constructs and their manifestations in individuals, leading to more accurate and reliable assessments. By focusing on traits with real existence, researchers can enhance the scientific rigor of their assessments and avoid ad hoc reasoning, ultimately contributing to the systematic and scientific advancement of psychometrics.

Why is focusing on real traits important for assessment accuracy?

  • Focusing on real traits in assessments is crucial for ensuring accuracy and validity in psychological measurements. By prioritizing traits that have real existence independent of the measurement process, researchers can develop a deeper understanding of psychological constructs and their manifestations in individuals, leading to more reliable assessments. This approach helps in grounding the assessments in actual psychological phenomena, rather than relying on ad hoc reasoning or subjective judgments. Emphasizing real traits also contributes to the systematic and scientific advancement of psychometrics, as it aligns the assessments with objective psychological realities. Ultimately, by focusing on real traits, researchers can enhance the validity and reliability of their assessments, leading to more accurate and meaningful results in psychological testing.

Previous[05/02/2024] Reads on Construct ValidationNext[04/18/2024] AI for Assessment

Last updated 1 year ago