Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ

2.8K posts

Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ banner
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ

Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ

@GuyProchilo

Psychological Scientist | Lecturer of Research Methods | Fourth Year Psychology (Honours) Program Coordinator at @ISNPsych | https://t.co/UyDK4jyC0Q

Melbourne, Australia ๊ฐ€์ž…์ผ Ocak 2013
192 ํŒ”๋กœ์ž‰2.9K ํŒ”๋กœ์›Œ
๊ณ ์ •๋œ ํŠธ์œ—
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
Many power analyses I read in published manuscripts are not reported with sufficient transparency to understand what on earth they are doing! What is the effect estimate? Where did it come from? For what test does it apply? Etc. Here is a reporting guide I give students #phdchat
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ tweet media
English
18
204
817
0
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
note: all these images are just random screenshots of papers where the methods section makes claims about Little's test that it doesn't actually provide. Some even claim it provides support for MAR...
English
0
0
1
275
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
To sum it up: Missing data is complicated. Littleโ€™s test probably isnโ€™t doing what you think it is, and MCAR is often unlikely. Donโ€™t rely on Little's test as an excuse to avoid properly handling your missing data.
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ tweet media
English
1
0
0
359
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
Missing at Random (MAR) means that the probability of data being missing is purely random after conditioning on observed data (e.g., in multiple imputation). That is to say, it depends only on the observed data, not the missing values themselves. #phdchat #statstwitter
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ tweet media
English
1
0
5
865
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
Just wrapped up today's lecture on missing data! Covered item- and construct-level missingness, detection, mechanisms, analysis (using Welch's t-tests with missingness indicators), solutions, and a theory intro to multiple imputation (since #jamovi lacks modules). #phdchat
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ tweet media
English
0
0
3
464
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
Don't use a significance test of heterogeneity (eg Cochran's Q) to choose between fixed or random effects meta-analysis. If studies estimate a single common effect size, use fixed effects. If studies differ in populations, measures, or protocols, opt for random effects. #phdchat
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ tweet media
English
1
0
2
348
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
A fixed effects meta-analysis assumes a single common effect size across all studies, which is often unrealistic given differences in interventions, populations, and measurements. In psychology especially, a random effects meta-analysis is usually the more plausible #phdchat
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ tweet media
English
2
4
22
2.5K
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
Legha, A., Riley, R. D., Ensor, J., Snell, K. I. E., Morris, T. P., & Burke, D. L. (2018). Individual participant data metaโ€analysis of continuous outcomes: A comparison of approaches for specifying and estimating oneโ€stage models. Statistics in Medicine, 37(29), 4404โ€“4420.
English
0
0
3
319
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ
A one-stage IPD meta-analysis aggregates data across all studies and can be analyzed using a linear mixed model with the study ID as a random effect. This aligns with a random effects meta-analysis where we assume the true effect size differs across studies #phdchat #statstwitter
Dr. Guy Prochilo ๐Ÿณ๏ธโ€๐ŸŒˆ tweet media
English
1
4
35
10.9K