Shachar Hochman's Avatar

Shachar Hochman

@hochmanshachar

Data Enthusiastic • Researcher • Applied Statistician Ex-Academic (Cognitive Psychology, PhD) My blog: https://cogpsychreserve.netlify.app My LinkedIn: https://www.linkedin.com/in/shachar-hochman-phd/

124
Followers
162
Following
22
Posts
15.08.2023
Joined
Posts Following

Latest posts by Shachar Hochman @hochmanshachar

Post image

1/9 New blog is live! This is part 2 of a series—last time we looked at the Dunning-Kruger effect, now we are digging in to Implicit vs Explicit attitudes and the Implicit Association Test. To start, of course we need a good meme...

haines-lab.com/post/part-2-...

26.01.2026 17:45 👍 42 🔁 15 💬 4 📌 3
Preview
A novel task for measuring numerical bias among adults Numerical bias is the spontaneous tendency to base decisions on numerical rather than equally available non-numerical information. We introduce the Co…

📄 Paper: www.sciencedirect.com/science/arti...

#AcademicSky #CognitiveScience #Psychology #NumericalCognition #MathCognition #IndividualDifferences #Psychometrics #Bayesian #OpenScience #OpenData #NewPaper

10.01.2026 13:14 👍 4 🔁 1 💬 1 📌 0

This has been a long journey. It started in April 2019 with a real eureka moment. I feel incredibly lucky to have @mattansb.msbstats.info as a major wind and support (and friend!) throughout this work. Years of work with Roi Cohen Kadosh and Avishai Henik, and we're thrilled it's finally out.

10.01.2026 13:14 👍 2 🔁 0 💬 1 📌 0

We also used a numerical Stroop task and found that stronger numerical bias in CLIP predicts larger Stroop effects when numbers must be ignored. This suggests a "tendency layer": varying attentional processes that predetermine how hard it is to ignore irrelevant information.

10.01.2026 13:14 👍 2 🔁 0 💬 1 📌 0
Post image

These individual differences predicted math fluency and quantitative reasoning, echoing child SFON research.

10.01.2026 13:14 👍 2 🔁 0 💬 1 📌 0
Post image

Hierarchical Bayesian Drift Diffusion modeling lets us combine choice and RT into a single measure, avoiding reliability problems of difference scores. The result: fantastic internal reliability and stable individual differences across all CLIP conditions.

10.01.2026 13:14 👍 2 🔁 0 💬 1 📌 0

Why this matters: Research on children shows that spontaneous focus on numerosity (SFON) predicts math skills and math development. The CLIP task provides a computerized version that can fit both adults and children! It captures these spontaneous tendencies trial by trial, and works great online!

10.01.2026 13:14 👍 2 🔁 0 💬 1 📌 0
Post image

Our paper is finally out in Cognition! 🎉
We introduce the "CLIP task"—a computerized paradigm for measuring numerical bias in adults: when number and physical size both matter, do you spontaneously rely more on numbers or on physical size?

10.01.2026 13:14 👍 13 🔁 5 💬 1 📌 2

This has been a long journey. It started in April 2019 with a real eureka moment. I feel incredibly lucky to have @mattansb.msbstats.info as a major wind and support throughout this work. Years of work with Roi Cohen Kadosh and Avishai Henik, and we're thrilled it's finally out.

10.01.2026 12:56 👍 0 🔁 0 💬 0 📌 0
Post image

These individual differences predicted math fluency and quantitative reasoning, echoing child SFON research.

10.01.2026 12:56 👍 0 🔁 0 💬 0 📌 0
Post image

Hierarchical Bayesian Drift Diffusion modeling lets us combine choice and RT into a single measure, avoiding the reliability problems of difference scores. The result: fantastic internal reliability and stable individual differences across all CLIP conditions.

10.01.2026 12:56 👍 0 🔁 0 💬 1 📌 0

Why this matters: Research on children shows that spontaneous focus on numerosity (SFON) predicts math skills and math development. The CLIP task provides a computerized version that can fit both adults and children! It captures these spontaneous tendencies trial by trial, and works great online!

10.01.2026 12:56 👍 0 🔁 0 💬 1 📌 0
OSF

New preprint with @rogierk.bsky.social @paulbuerkner.com - we introduce "relative measurement uncertainty" - a reliability estimation method that's applicable across a broad class of Bayesian measurement models (e.g., generative-, computational- and item response theory-models osf.io/h54k8

01.10.2025 08:17 👍 19 🔁 7 💬 2 📌 5
Preview
Exploring {ggplot2}’s Geoms and Stats – Stat’s What It’s All About

New blog post!

Ever wonder what geom_histogram is actually doing? How about geom_boxplot?

In celebration of the release of #ggplot2 4.0.0 (ggplot8?), I explore the relationships between the “geoms” and “stats” offered by the core {ggplot2} functions.

#rstats

15.09.2025 19:04 👍 80 🔁 35 💬 1 📌 4
Preview
Within-person factorial experiments, log(normal) reaction-time data | A. Solomon Kurz Causal inference with the GLMM, Part 1

New #rstats blog up!

solomonkurz.netlify.app/blog/2025-07...

This is the first in a new series discussing causal inference with experimental data using multilevel models. My basic case is g-computation is the way to go.

21.07.2025 14:14 👍 97 🔁 19 💬 7 📌 1
Beyond the Exclamation Points!!! – CogPsych Reserve

Dive in for code, visuals, and a clearer path through the log-odds fog → cogpsychreserve.netlify.app/posts/logist...

#NLP #Kaggle #marginaleffects #BayesianStatistics #DataScience #SignificantTesting

14.07.2025 07:14 👍 5 🔁 1 💬 0 📌 0

2/3

• NLP + PCA to capture toxicity/incoherence
• Cohen’s d ➡️ log-odds priors in one line using #brms
• #marginaleffects → 0–100 % probability shifts you can explain
• Inference with HDI-ROPE. It flags which effects are big enough to matter. Great for researchers and anyone shipping spam filters!

14.07.2025 07:14 👍 2 🔁 0 💬 1 📌 0
Post image

1/3 New post up! 📝 I took the workhorse 🔧 of binary modeling—logistic regression—and gave it a Bayesian tune-up using a Kaggle SMS-spam dataset.

14.07.2025 07:14 👍 19 🔁 6 💬 2 📌 2

Thanks Laura! 🙏 I analyzed vertical-face tasks (6 variants across SOAs) from subjects with mouse responses only. The Preprocessing details are in the post’s collapsible section 😊. Grateful for your work—DM anytime!

08.03.2025 20:22 👍 3 🔁 0 💬 0 📌 0

Thank you! 😊 While latent correlations are possible in Stan via custom likelihoods (modeling latent Gaussian variables), it's quite involved. For 95% of cases, I recommend the simpler brms approach: model questionnaires as predictors of task effects using condition-by-questionnaire interactions.

08.03.2025 20:05 👍 2 🔁 0 💬 0 📌 0

6/6 Thanks to @solomonkurz.bsky.social for statistical inspiration, @natehaines.bsky.social for works that influenced my approach, and @almogsi.bsky.social & @mattansb.bsky.social or thoughtful feedback!

#BayesianStatistics #ReliabilityAnalysis #CognitiveScience

07.03.2025 09:14 👍 5 🔁 0 💬 0 📌 0

5/6 The implications go beyond this single task. Many measures in psychology (and beyond) might be more reliable than we thought—we need to preserve and properly model the information in trial-level data.

07.03.2025 09:14 👍 3 🔁 0 💬 1 📌 0
Post image

4/6 This visualization shows the transformation when the same data is analyzed with trial-level Bayesian methods instead of traditional aggregation:

07.03.2025 09:14 👍 4 🔁 0 💬 1 📌 0

3/6 I implemented two Bayesian approaches in #brms:

@jeffrouder.bsky.social & @juliaha.bsky.social's variance decomposition
@gangchen6.bsky.social's approach

Both show substantially higher reliability than traditional analyses.

07.03.2025 09:14 👍 3 🔁 0 💬 1 📌 0

2/6 Recent research by @irenexu.bsky.social claimed the emotional dot-probe task lacks reliability for individual differences research. I wanted to see if more sophisticated analysis methods could tell a different story.

07.03.2025 09:14 👍 2 🔁 0 💬 1 📌 0
The Dot-Probe Task is Probably Fine – CogPsych Reserve

1/6 Hello Bluesky! 👋 Excited to join this community and share my new blog. First post: Using Bayesian hierarchical models to rescue "unreliable" cognitive tasks, with the dot-probe task as my case study. cogpsychreserve.netlify.app/posts/dotpro...

07.03.2025 09:14 👍 64 🔁 17 💬 7 📌 8