I checked 15 psychology journals on Tuesday, January 27, 2026 using the Crossref API. For the period January 20 to January 26, I found 28 new paper(s) in 9 journal(s).

Behavior Research Methods

The point of subjective equality as a tool for accurate and robust analysis in categorization tasks
Ariel Levy, Tali Kleiman, Yuval Hart
Full text
Ten particularly frequent and consequential questionable research practices in quantitative research: Bias mechanisms, preventive strategies, and a simulation-based framework
Theodoros A. Kyriazos, Mary Poga
Full text
Using large language models to estimate belief strength in reasoning
Jérémie Beucler, Zoe Purcell, Lucie Charles, Wim De Neys
Full text
Test–retest reliability of the gaze index for sign-tracking and goal-tracking
Marco Badioli, Claudio Danti, Luigi Degni, Gianluca Finotti, Valentina Bernardi, Lorenzo Mattioni, Francesca Starita, Giuseppe di Pellegrino, Sara Giovagnoli, Mariagrazia Benassi, Sara Garofalo
Full text
Constructing the Corpus of Children’s Video Media (CCVM): A new resource and guidelines for constructing comparable and reusable corpora
Anna Gowenlock, Jennifer Rodd, Beth Malory, Courtenay Norbury
Full text
A growing number of psycholinguistic studies use methods from corpus linguistics to examine the language that children encounter in their environment, to understand how they might acquire different aspects of linguistic knowledge. Many of these studies focus on child-directed speech or children’s literature, while there is a paucity of work focusing on children’s television and video media. We describe the creation and contents of the Corpus of Children’s Video Media (CCVM), a specialised corpus designed to represent the spoken language in television and online videos popular among 3–5-year-old children in the UK (available as a scrambled database of tokens). The CCVM was designed to be comparable to an existing corpus of child-directed speech (CDS). We used a dual sampling approach: inclusion decisions were guided by (a) a survey of parents with children in our target age group, and (b) a survey of programmes available on popular streaming platforms. The corpus consists of 233,471 tokens across 161 transcripts (43.12 h of video) and is available on the Open Science Framework (OSF) as a scrambled database of tokens (including gloss, stem, and lemma forms, and part-of-speech tags), organised within transcripts, together with relevant metadata for each transcript. We discuss the challenges of creating a corpus that is comparable to existing datasets and highlight the importance of transparency in this process. We take an open science approach, sharing a detailed data collection and processing protocol, code, and data so that the corpus can be evaluated, extended, and used appropriately by other research teams.
Modeling truncated and censored data with the diffusion model in Stan
Franziska Henrich, Karl Christoph Klauer
Full text
Reaction time data in psychology are frequently censored or truncated. For example, two-alternative forced-choice tasks that are implemented with a response window or response deadline give rise to censored or truncated data. This must be accounted for in the data analysis, as important characteristics of the data, such as the mean, standard deviation, skewness, and correlations, can be strongly affected by censoring or truncation. In this paper, we use the probabilistic programming language Stan to analyze such data with Bayesian diffusion models. For this purpose, we added the functionality to model truncated and censored data with the diffusion model by adding the cumulative distribution function for reaction times generated from the diffusion model and its complement to the source code of Stan. We describe the usage of the truncated and censored models in Stan, test their performance in recovery and simulation-based calibration, and reanalyze existing datasets with the new method. The results of the recovery studies are satisfactory in terms of correlations ( $$r=.93 - 1.00$$ r = . 93 - 1.00 ), coverage (93–95% of true values lie in the 95% highest density interval), and bias. Simulation-based calibration studies suggest that the new functionality is implemented without errors. The reanalysis of existing datasets further validates the new method.
Jokers in the deck: A new temperature setting for the Columbia Card Task
Kevin Kapadia, Yunxiu Tang, Richard John
Full text
The Columbia Card Task (CCT) is a behavioral measure of risk-taking (BMRT), which has been cited over 1,500 times (Google Scholar, 3/1/2024). The original game had two versions (Hot and Cold), measuring affective and deliberative decision-making, respectively. Each version included 54 scored rounds where the loss cards were placed at the end, and nine unscored rounds where the loss cards were placed systematically among the gain cards. Over time, the CCT has gone through many iterations on critical components, such as the number of rounds, the position of the loss cards, and the introduction of a new version (Warm). Despite this, there are several issues with the CCT, notably a need for convergent validity with other measures of risk-taking. This paper reviews different iterations of the CCT, introduces a new (Toasty) version of the CCT that is a hybrid of the hot and warm versions, explores the consequences of randomly placing the loss cards among the gain cards consistent with instructions provided to participants, and examines the impact of incentivizing participants based on their score. Results ( N = 405) show that the Toasty version behaves similarly to the Warm but provides additional insights into risk-taking behavior. When loss cards are placed randomly, participants are still sensitive to the game’s parameters (gain amount, loss amount, and number of loss cards) and reveal the loss cards roughly half the time. Incentivizing participants in our study had little impact on the number of cards revealed.
Accuracy in parameter estimation and simulation approaches for sample-size planning accounting for item effects
Erin M. Buchanan, Mahmoud M. Elsherif, Jason Geller, Chris L. Aberson, Necdet Gurkan, Ettore Ambrosini, Tom Heyman, Maria Montefinese, Wolf Vanpaemel, Krystian Barzykowski, Carlota Batres, Katharina Fellnhofer, Guanxiong Huang, Joseph McFall, Gianni Ribeiro, Jan P. Röer, José L. Ulloa, Timo B. Roettger, K. D. Valentine, Antonino Visalli, Kathleen Schmidt, Martin R. Vasilev, Giada Viviani, Jacob F. Miranda, Savannah C. Lewis
Full text
The planning of sample size for research studies often focuses on obtaining a significant result given a specified level of power, significance, and an anticipated effect size. This planning requires prior knowledge of the study design and a statistical analysis to calculate the proposed sample size. However, there may not be one specific testable analysis from which to derive power (Silberzahn et al., Advances in Methods and Practices in Psychological Science , 1 (3), 337356, 2018) or a hypothesis to test for the project (e.g., creation of a stimuli database). Modern power and sample size planning suggestions include accuracy in parameter estimation (AIPE, Kelley, Behavior Research Methods , 39 (4), 755–766, 2007; Maxell et al., Annual Review of Psychology , 59 , 537–563, 2008) and simulation of proposed analyses (Chalmers & Adkins, The Quantitative Methods for Psychology , 16 (4), 248–280, 2020). These toolkits offer flexibility in traditional power analyses that focus on the if-this, then-that approach. However, both AIPE and simulation require either a specific parameter (e.g., mean, effect size, etc.) or a statistical test for planning sample size. In this tutorial, we explore how AIPE and simulation approaches can be combined to accommodate studies that may not have a specific hypothesis test or wish to account for the potential of a multiverse of analyses. Specifically, we focus on studies that use multiple items and suggest that sample sizes can be planned to measure those items adequately and precisely, regardless of the statistical test. This tutorial also provides multiple code vignettes and package functionality that researchers can adapt and apply to their own measures.
Measuring individual differences in the speed of attention using the distractor intrusion task
Alon Zivony, Claudia C. von Bastian, Rachel Pye
Full text
How quickly we attend to objects plays an important role in navigating the world, especially in dynamic and rapidly changing environments. Measuring individual differences in attention speed is therefore an important, yet challenging, task. Although reaction times in visual search tasks have often been used as an intuitive proxy of such individual differences, these measures are limited by inconsistent levels of reliability and contamination by non-attentional factors. This study introduces the rate of post-target distractor intrusions (DI) in the rapid serial visual presentation (RSVP) paradigm as an alternative method of studying individual differences in the speed of attention. In RSVP, a target is presented for a brief duration and embedded among multiple distractors. DIs are reports of a subsequent distractor rather than the target and have previously been shown to be associated with the speed of attention. The present study explored the reliability and validity of DI rates as a measure of individual differences. In three studies, DI rates showed high internal consistency and test–retest reliability over a year (>.90), even with a short task administration of only about 5 minutes. Moreover, DI rates were associated with measures related to attention speed, but not with unrelated measures of attentional control, reading speed, and attentional blink effects. Taken together, DI rates can serve as a useful tool for research into individual differences in the speed of attention. Links to a downloadable and easily executable DI experiment, as well as a brief discussion of methodological considerations, are provided to facilitate such future research.

Computers in Human Behavior

Training and oversight of algorithms in social decision-making: Algorithms with prescribed selfish defaults breed selfish decisions
Terence D. Dores Cruz, Mateus A.M. de Lucena
Full text
User Cognitive Fit in Human-AI Interaction: Exploring the Link Between Input Representation and Generative Output Complexity
Hechang Cai, Jinlai Zhou
Full text
Brain Responses to Deepfakes and Real Videos of Emotional Facial Expressions Reveal Detection Without Awareness
Casey Becker, Russell Conduit, Philippe A. Chouinard, Robin Laycock
Full text
Evaluating Human-Based and Large Language Model-Based Content Analysis: A Case Study on Negative Word-of-Mouth Classification in Social Media
Yolande Yang, Chih-Chien Wang, Pau-Lin Chou, Chieh-Yu Tien, Jia-Ci Wen, Nien-Hsin Chen
Full text

Group Processes & Intergroup Relations

The effect of common humanity on forgiveness and collective action does not generalise to all victim groups
Ă–zden Melis UluÄź, Yasemin GĂĽlsĂĽm Acar, Katharine H. Greenaway, Brian Lickel, Ă–bek Tutku EroÄźlu
Full text
Prior work has found that making common human identity salient increases forgiveness of perpetrators but reduces collective action intentions among victims of historical atrocities. We conducted three experiments to investigate the generalisability of this effect among Alevis—a religious minority found mainly in Turkey. Study 1 ( N = 222) found that the common human identity manipulation was unsuccessful and did not lead to differences in forgiveness of perpetrators or collective action intentions among Alevis. In Study 2 ( N = 164), we conducted the same experiment and asked an open-ended question about norms associated with Alevism. The results replicated the null effects in Study 1, additionally showing that the social norms of Alevis relate to being humanist and peaceful. Study 3 ( N = 183) tested the role of Alevis’ humanist norms as a potential moderator of the inclusiveness effect. We again found that the common human identity manipulation did not affect forgiveness or collective action intentions, nor did Alevism norms moderate this effect. We discuss the importance of generalising social psychological findings to different cultural contexts and different victim groups.

Journal of Experimental Social Psychology

Generic title: Not a research article
Corrigendum
Zhihe Pan, Hweemin Tan, Siqi Liu, Xia Fang
Full text
Status decoded: How actors and observers shape the meaning of stealth symbols
Jesse D'Agostino, Derek D. Rucker
Full text
Proximity to whiteness and the racial position of multiracial people in the United States
A. Chyei Vinluan, Maria G. Garay, Jennifer M. Perry, Linda X. Zou, Keith B. Maddox, Jessica D. Remedios
Full text

Journal of Personality and Social Psychology

“Why didn’t you just say so?” People use indirect opposition to assess partner commitment.
Levi R. Baker, James K. McNulty, V. Michelle Russell
Full text
Anxious aspirations: Attachment anxiety fuels status strivings through intrasexual competition.
Agata Gasiorowska, Michał Folwarczny, Tobias Otterbring
Full text
Using machine learning to predict individual differences in psychological reactivities to social interactions.
Ole Hätscher, Johannes L. Klinz, Niclas Kuper, Lara Kroencke, Julian Scharbert, Eric Grunenberg, Mitja D. Back
Full text

Multivariate Behavioral Research

Multiple Imputation of Missing Data in Moderated Factor Analysis
Joost R van Ginkel, Dylan Molenaar
Full text
Time-Varying Path-Specific Direct and Indirect Effects: A Novel Approach to Examine Dynamic Behavioral Processes with Application to Smoking Cessation
Yajnaseni Chakraborti, Recai M. Yucel, Megan E. Piper, Jeremy Mennis, Anthony J. Alberg, Timothy B. Baker, Donna L. Coffman
Full text
Analyzing Complex Treatment Effects in Nonrandomized Observational Studies: The Case of Retention of Students in Grade
Stephen G. West
Full text

Personality and Social Psychology Bulletin

The Dialectical Self Around the World: A Meta-Analysis of Country-Level Means
Julie Spencer-Rodgers, Isabella Major-Siciliano, Wei Yan, Antonio A.S. Cortijo, Lauren McKenzie, Kaiping Peng
Full text
This research presents the first known meta-analysis of the Dialectical Self Scale, a widely used measure of the extent to which people hold contradictory and changeable self-conceptions. Data were synthesized from k = 139 studies ( N = 23,629) from 28 countries to produce a national Dialectical Self Index (DSI). Study 1 used meta-analytic techniques to hierarchically order countries on dialecticism and test demographic moderators. No historical shifts in dialecticism were observed over two decades. In Study 2, dialecticism, at the country-level, was correlated with variables reflecting tolerance of contradiction and expectation of change, and socioecological factors (Buddhism, rice farming), but only weakly related to contemporary macro-social forces (globalization). Dialecticism was unrelated to collectivism and interdependent self-construals, indicating it is a foundational cultural mindset. A world map of dialecticism showed clear regional clustering. The DSI provides a useful tool for conducting cross-national research on dialecticism.
Changing Norms Following the 2024 U.S. Presidential Election: The Trump Effect on Prejudice Redux
Samuel E. Arnold, Jenniffer Wong Chavez, Kelly S. Swanson, Christian S. Crandall
Full text
Following the 2016 U.S. Presidential election of Donald Trump, prejudice toward groups targeted during his campaign (e.g., Asian Americans, Mexicans) become more acceptable. By contrast, both Trump and Clinton voters reported less prejudice of their own. We conducted a 2024 conceptual replication, measuring perceived norms of prejudice and own-prejudice toward 128 groups, both before ( N = 362) and after ( N = 261) the U.S. election. We separately measured the negativity of Trump’s campaign rhetoric toward these groups ( N = 188). Levels of prejudice and perceived norms of prejudice acceptability were mostly stable pre-/post-election, but Trump’s negative rhetoric predicted an increase in perceived acceptability of prejudice among targeted groups (replicating the 2016 results), and a rise in self-reported prejudice in the same groups post-election (reversing the 2016 results). Despite changes in the sociopolitical context between elections, the election of a leading politician who campaigned on prejudice was again associated with increases in the acceptability of prejudice.

Psychological Methods

Testing and improving the robustness of amortized bayesian inference for cognitive models.
Yufei Wu, Stefan T. Radev, Francis Tuerlinckx
Full text

Psychology of Popular Media

This is why they are called “social” media! Nonlinear associations between time spent on social media and psychosocial well-being in a representative Italian sample.
Tommaso Galeotti, Natale Canale, Claudia Marino, Michela Lenzi, Massimiliano Pastore, Michela Bersia, Daniela Pierannunzio, Giacomo Lazzeri, Alessio Vieno
Full text
Unmasking entities: Expectancy violations in VTubers’ and virtual influencers’ self-disclosure.
Yingjia Huang, Rui Gu
Full text
Do the traits of fictional crushes and real-life ideals align? An investigation into ideal standards based on fictional fandoms.
Nicole Zhi Min Wang, Ai Ni Teoh
Full text