Wednesday, April 23, 2025

On #factoranalysis of #IQ tests—impact of software choice—plus comments about art+science of factor analysis in #intelligence test research—#schoolpsychology



Contributing to the reproducibility crisis in Psychology: The role of statistical software choice on factor analysis.  Journal of School Psychology.  Stefan C. Dombrowski.  Click here to view article source and abstract.

This is an important article for those who conduct (and also those who consume) factor analysis results of intelligence or cognitive ability tests.  

Abstract (note - bold font in abstract has been added by me)

A potentially overlooked contributor to the reproducibility crisis in psychology is the choice of statistical application software used for factor analysis. Although the open science movement promotes transparency by advocating for open access to data and statistical methods, this approach alone is insufficient to address the reproducibility crisis. It is commonly assumed that different statistical software applications produce equivalent results when conducting the same statistical analysis. However, this is not necessarily the case. Statistical programs often yield disparate outcomes, even when using identical data and factor analytic procedures, which can lead to inconsistent interpretation of results. This study examines this phenomenon by conducting exploratory factor analyses on two tests of cognitive ability—the WISC-V and the MEZURE—using four different statistical programs/applications. Factor analysis plays a critical role in determining the underlying theory of cognitive ability instruments, and guides how those instruments should be scored and interpreted. However, psychology is grappling with a reproducibility crisis in this area, as independent researchers and test publishers frequently report divergent factor analytic results. The outcome of this study revealed significant variations in structural outcomes among the statistical software programs/applications. These findings highlight the importance of using multiple statistical programs, ensuring transparency with analysis code, and recognizing the potential for varied outcomes when interpreting results from factor analytic procedures. Addressing these issues is important for advancing scientific integrity and mitigating the reproducibility crisis in psychology particularly in relation to cognitive ability structural validity.

My additional comments

The recommendation that multiple factor analysis software programs be used when analyzing the structural validity of cognitive abilities tests makes sense.  Kudos to Dr. Dombrowski for demonstrating this need.

Along these lines, it is also important to recognize that the use and interpretation of any factor analysis software is highly dependent on the statistical and substantive expertise and skills of the researcher.  I made these points (based on the writings and personal conversations with Jack Carroll) in a recent article (McGrew, 2023; open access so you can download and read) in the Journal of Intelligence.  The salient material is reproduced below.  This article can be accessed either a the journal website or via the Research and Reports section of my MindHub web page (McGrew, 2023)


(Note - Bold font in text below, extracted from McGrew (2023), is not in the original published article)

“I was fortunate to learn important tacit EFA and CFA knowledge during my 17 years of interactions with Carroll, and particularly my private one-to-one tutelage with Carroll in May 2003. Anyone who reads Chapter 3 (Survey and Analysis of Correlational and Factor-Analytic Research on Cognitive Abilities: Methodology) of Carroll's 1993 book, as well as his self-critique of his seminal work (Carroll 1998) and other select method-focused post-1993 publications (Carroll 1995, 1997), should conclude what is obvious—to Carroll, factor analyses were a blend of art and science. As articulated by some of his peers (see footnote #2), his research reflected the work of an expert with broad and deep substantive knowledge of research and theories in intelligence, cognitive psychology, and factor analysis methods. 

In 2003, after Carroll had been using CFA to augment his initial EFA analyses for at least a decade, Carroll expressed (to me during our May 2003 work week) that he was often concerned with the quality of some reported factor analyses (both EFA and CFA) of popular clinical IQ tests or other collections of cognitive ability measures (Carroll 1978, 1991, 1995, 2003). Carroll's characteristic positive skepticism regarding certain reported factor analyses was first articulated (as far as I know) in the late 1970's, when he stated “despite its many virtues, factor analysis is a very tricky technique; in some ways it depends more on art than science, that is, more on intuition and judgment than on formal rules of procedure. People who do factor analysis by uncritical use of programs in computer packages run the risk of making fools of themselves” (Carroll 1978, p. 91; emphasis added). It is my opinion that Carroll would still be dismayed by some of the EFA and CFA studies of intelligence tests published during the past two decades that often used narrow or restricted forms of factor analysis methods and rigid formal statistical rules for decision-making, with little attempt to integrate contemporary substantive research or theory to guide the analysis and interpretation of the results (e.g., see Decker 2021; Decker et al. 2021; McGrew et al. 2023). 

Carroll's unease was prescient of recently articulated concerns regarding two aspects of the theory crises in structural psychological research—the conflation of statistical (primarily factor analysis) models with theoretical models and the use of narrow forms of factor analysis methods (Fried 2020; McGrew et al. 2023). First, many intelligence test batteries only report CFA studies in their technical manuals. EFA results, which often produce findings that vary from CFA findings, are frequently omitted. This often leads to debates between independent researchers and test authors (or test publishers) regarding the validity of the interpretation of composite or cluster scores, leaving test users confused regarding the psychometric integrity of composite score interpretations. McGrew et al. (2023) recently recommended that intelligence test manuals, as well as research reports by independent researchers, include both EFA and CFA (viz., bifactor g, hierarchical g, and Horn no-g models), as well as psychometric network analysis (PNA) and possibly multidimensional scaling analyses (MDSs; McGrew et al. 2014; Meyer and Reynolds 2022). As stated by McGrew et al. (2023), “such an ecumenical approach would require researchers to present results from the major classes of IQ test structural research methods (including PNA) and clearly articulate the theoretical basis for the model(s) the author's support. Such an approach would also gently nudge IQ test structural researchers to minimize the frequent conflation of theoretical and psychometric g constructs. Such multiple-methods research in test manuals and journal publications can better inform users of the strengths and limitations of IQ test interpretations based on whatever conceptualization of psychometric general intelligence (including models with no such construct) underlies each type of dimensional analysis” (p. 24).”


Friday, April 18, 2025

#ITC #webinar: Test #adaptation reporting #standards - May 8, 2025

 Additional information regarding the Interntional Test Commission (ITC) and registration info can be found here.


Click on image to enlarge.

Research Byte: Understanding #Adaptive Skills in #Borderline #IntellectualFunctioning: A Systematic #Review - #schoolpsychology #DD #intelligence #developmentaldisabilities #SPED


 Click on image to enlarge for easy reading

Understanding Adaptive Skills in Borderline Intellectual Functioning: A Systematic Review

Open access copy available by clicking here.

Abstract

Borderline intellectual functioning (BIF) is characterized by an IQ typically ranging from 70 to 85, combined with deficits in adaptive functioning. Despite its prevalence, individuals with BIF are often excluded from diagnostic and support systems, which traditionally focus on strictly defined intellectual disabilities. This article presents a systematic review conducted across the ProQuest, WoS, SCOPUS, and EBSCOhost databases, aiming to develop a profile of the adaptive functioning in individuals with BIF. A total of 64 documents published from 2012 to the present were included, all of them addressing BIF and adaptive functioning skills, and quality was assessed using the SSAHS tool. The findings presented are synthesized according to conceptual, social, and practical domains and reveal that individuals with BIF experience widespread difficulties across the conceptual, social, and practical domains. Additionally, societal barriers, primarily limiting access to support services, persist. However, there are emerging resources aimed at supporting this population, such as legislative efforts to facilitate their integration into the labor market. The implications and limitations of the findings are discussed, highlighting the need to consider the adaptive functioning skills of individuals with BIF. 


Wednesday, April 09, 2025

Research Byte: Development of #Arithmetic Across the #Lifespan: A Registered Report. - #Gq #CHC #Gwm #EF #Gs #schoolpsychology #SPED #SLD


Click on image to enlarge for easy viewing

Development of Arithmetic Across the Lifespan: A Registered Report.  


Open access paper available at Developmental Psychology journal.  Click here to access

Abstract
 
Arithmetic skills are needed at any age. In everyday life, children to older adults calculate and deal with numbers. The processes underlying arithmetic seem to change with age. From childhood to younger adulthood, children get better in domain-specific numerical skills such as place-value processing. From younger to older adulthood, domain-general cognitive skills such as working memory decline. These skills are needed for complex arithmetic such as addition with carrying and subtraction with borrowing. This study investigates how the domain-specific (number magnitude, place-value processing) and domain-general (working memory, processing speed, inhibition) processes of arithmetic change across the lifespan. Thereby, arithmetic effects (carry and borrow effects), numerical effects (distance and compatibility effects), and cognitive skills were assessed in children, younger and older adolescents, and younger, middle-aged and older adults. The results showed that numerical and arithmetic skills improve from childhood to young adulthood and remain relatively stable throughout adulthood, even though domain-general pro-cesses, particularly working memory and processing speed, decline with age. While number magnitude and place-value processing both develop until adulthood, number magnitude processing shows deficits during aging, whereas place-value processing remains intact even in old age. The carry effect shifts from a categorical all-or-none decision (whether or not a carry operation is needed) to a more continuous magnitude process in adulthood, reflecting increasing reliance on domain-specific skills. In contrast, the borrow effect remains largely categorical across all age groups, depending on general cognitive processes. These results provide critical insights into how arithmetic skills change over the lifespan, relying on both domain-specific and domain-general processes.

Public Significance Statement 

Numerical and arithmetic skills improve significantly during school and are mostly preserved throughout adulthood—despite a decline in cognitive skills such as working memory and processing speed during aging. When facing complex arithmetic, all—from children up to older adults—need longer to calculate, but lifelong experience helps in dealing with arithmetic complexity. Throughout the lifespan, arithmetic requires both cognitive skills as well as numeric skills.

Tuesday, April 08, 2025

Research Byte: Conjectures and refutations in #cognitive ability #structuralvalidity research [with #WISC-V]: Insights from Bayesian structural equation modeling - #schoolpsychology #IQ #intelligence #Wechslers #WISC-V

Conjectures and refutations in cognitive ability structural validity research [with #WISC-V]: Insights from Bayesian structural equation modeling

Click here to view Journal of Psychology of School Psychology source of publication - not open access.

Abstract

The use of Bayesian structural equation modeling (BSEM) provided additional insight into the WISC–V theoretical structure beyond that offered by traditional factor analytic approaches (e.g., exploratory factor analysis and maximum likelihood confirmatory factor analysis) through the specification of all cross loadings and correlated residual terms. The results indicated that a five-factor higher-order model with a correlated residual between the Visual-Spatial and Fluid Reasoning group factors provided a superior fit to the four bifactor model that has been preferred in prior research. There were no other statistically significant correlated residual terms or cross loadings in the measurement model. The results further suggest that the WISC–V ten subtest primary battery readily attains simple structure and its index level scores may be interpreted as suggested in the WISC–V's scoring and interpretive manual. Moreover, BSEM may help to advance IQ theory by providing contemporary intelligence researchers with a novel tool to explore complex interrelationships among cognitive abilities—relationships that traditional structural equation modeling methods may overlook. It can also help attenuate the replication crises in school psychology within the area of cognitive assessment structural validity research through systematic evaluation of complex structural relationships obviating the need for CFA based post hoc specification searches which can be prone to confirmation bias and capitalization on chance.

Saturday, April 05, 2025

Research Byte: Reevaluating the #Flynneffect, and the reversal: Temporal trends and measurement invariance in Norwegian armed forces #intelligence scores

 

Reevaluating the Flynn effect, and the reversal: Temporal trends and measurement invariance in Norwegian armed forces intelligence scores

Open access PDF available from journal Intelligenceclick here.

Abstract

Since 1954, the Norwegian Armed Forces have annually administered an unchanged general mental ability test to male cohorts, comprising figure matrices, word similarities, and mathematical reasoning tests. These stable and representative data have supported various claims about shifts in general mental ability (GMA) levels, notably the Flynn effect and its reversal, influencing extensive research linking these scores with health and other outcomes. This study examines whether observed temporal trends in scores reflect changes in latent intelligence or are confounded by evolving test characteristics and specific test-taking abilities in numerical reasoning, word comprehension, and figure matrices reasoning. Our findings, using multiple-group factor analysis and multiple indicator multiple cause (MIMIC) models, indicate that while there was a general upward trend in observed scores until 1993, this was predominantly driven by enhancements in the fluid intelligence task, specifically figure matrices reasoning. Notably, these gains do not uniformly translate to a rise in underlying GMA, suggesting the presence of domain-specific improvements and test characteristic changes over time. Conversely, the observed decline is primarily due to decreases in word comprehension and numerical reasoning tests, also reflecting specific abilities not attributable to changes in the latent GMA factor. Our findings further challenge the validity of claims that changes in the general factor drive the Flynn effect and its reversal. Furthermore, they caution against using these scores for longitudinal studies without accounting for changes in test characteristics.

Thursday, April 03, 2025

Research Byte: A Systematic #Review of Theoretical Frameworks in #Reading and #Writing: Insights from JAAL (2015–2024)—#Grw #CHC #schoolpsychology #education #SPED




 

A Systematic Review of Theoretical Frameworks in Reading and Writing: Insights from JAAL (2015–2024) 

Click here to visit journal web page.  Not an open access article.

Abstract

This systematic review explored the theoretical frameworks used in reading and writing research published in the Journal of Adolescent and Adult Literacy (JAAL) between 2015 and 2024. Using the matrix method, we screened 126 articles for inclusion criteria and identified 20 studies that addressed both reading and writing in empirical research. We coded each article to determine which theories were explicitly or implicitly used, focusing on five frameworks frequently cited in literacy studies: sociocultural theory, reader response theory, motivation theory, social constructivism, and transactional theory. Results revealed sociocultural theory as the dominant lens, whereas other theories were mainly used as secondary frameworks and often cited implicitly. We also found limited integration across theories despite the acknowledged interdependence of reading and writing. These findings indicate a need for greater theoretical clarity and suggest that researchers and practitioners adopt more integrated approaches when investigating or teaching literacy. By highlighting the current state of theoretical application in JAAL, this review underscores opportunities to strengthen both research design and classroom practice through more explicit and multifaceted theoretical grounding.

Tuesday, April 01, 2025

National Academies Press downloadable PDF publicaiton on #cognitive #aging - #developmental #cognitive #cognition




 Click here for more info and ability to download a free copy.

Institute of Medicine. 2015. Cognitive Aging: Progress in Understanding and Opportunities for Action. Washington, DC: The National Academies Press. https://doi.org/10.17226/21693.

Research Byte: Science and Practice of Identifying #SpecificLearningDisabilities (#SLD): Kind Conversations About a Wicked Problem —#SPED #schoolpsychology #LD





Science and Practice of Identifying Specific Learning Disabilities: Kind Conversations About a Wicked Problem - Daniel B. Hajovsky, Kathrin E. Maki, Christopher R. Niileksela, Ryan J. McGill, 2025, Journal of Psychoeducational Assessment.  

Click here to visit journal page (unfortunately not an open access PDF).


Abstract
Although specific learning disabilities (SLD) represent the largest category for which school-age children receive special education services, the science and practice of SLD identification continues to evade consensus. Our goal is to bring together trainers and researchers with different perspectives on SLD identification to help spur a move toward a potential consensus, discuss agreements and disagreements on SLD identification in the field including amongst ourselves, and work toward productive discussion that may help move the field forward. We review essential conceptual questions that require greater scrutiny and thought to build a stronger understanding of SLD. We then discuss current assessment and identification practices, focusing on the not-so-controversial and the controversial issues in the field. Finally, we conclude with questions and considerations that challenge many of the established assumptions and systems currently in place. The aim of this article is to support constructive discussion on the topic of SLD that may have profound effects on the perennial issues the field continues to face.

Friday, March 21, 2025

#ITC webinar: #Validity and #Fairness Considerations in Innovative #Digital #Assessments."

When and how can you access this webinar?

The webinar is live online on Thursday, April 3rd, at 10 a.m. (New York, USA time, UTC-4). 



 

Research Byte: Co-Occurrence and Causality Among #ADHD, #Dyslexia, and #Dyscalculia - #SLD #schoolpsychology #sped #genetics #EDPSY

Co-Occurrence and Causality Among ADHD, Dyslexia, and Dyscalculia

Published in Psychological Science.  Click here to access PDF copy of article

Abstract
ADHD, dyslexia, and dyscalculia often co-occur, and the underlying continuous traits are correlated (ADHD symptoms, reading, spelling, and math skills). This may be explained by trait-to-trait causal effects, shared genetic and environmental factors, or both. We studied a sample of ≤ 19,125 twin children and 2,150 siblings from the Netherlands Twin Register, assessed at ages 7 and 10. Children with a condition, compared to those without that condition, were 2.1 to 3.1 times more likely to have a second condition. Still, most children (77.3%) with ADHD, dyslexia, or dyscalculia had just one condition. Cross-lagged modeling suggested that reading causally influences spelling (β = 0.44). For all other trait combinations, cross-lagged modeling suggested that the trait correlations are attributable to genetic influences common to all traits, rather than causal influences. Thus, ADHD, dyslexia, and dyscalculia seem to co-occur because of correlated genetic risks, rather than causality.



 

Tuesday, March 11, 2025

Research Byte: Performance- and report-based measures of #executivefunction as predictors of children’s #academic skills - #neuropsychology #schoolpsychology



DeCamp, C., Alfonso, S. V., & Lonigan, C. J. (2025). Performance- and report-based measures of executive function as predictors of children’s academic skills. Neuropsychology, 39(3), 214–222. https://doi.org/10.1037/neu0000992

Abstract

Objective: Executive function (EF) is thought to be a core component of various cognitive processes. Two common ways to measure EF are through report-based measures that assess EF by collecting informant(s) reports on children’s behaviors and performance-based measures that assess EF through the completion of a task related to EF dimension(s). However, most research reports low associations between these measures. The goal of this study was to determine the unique and overlapping contributions of a report- and a performance-based measure of EF on children’s academic outcomes over time. Method: The sample consisted of 1,152 children (636 boys, 516 girls) who were part of a large-scale preschool intervention study. Children completed measures of academic achievement in kindergarten, first grade, and second grade, and they completed a performance-based measure of EF in kindergarten; teachers reported on children’s EF during the fall of kindergarten. Structural growth modeling was utilized to determine the unique and shared contributions of EF measures on concurrent ability and growth of academic outcomes. Results: Structural growth models indicated that the separate EF measures were both significant predictors of concurrent ability and growth of all academic outcomes, with one exception; the Head–Toes–Knees–Shoulders task was not a significant predictor of growth in math skills. Conclusions: Results of this study suggested that report- and performance-based measures of EF should not be used interchangeably, and these findings have implications for the utility of EF as a risk factor for poor academic achievement.