this post was submitted on 19 Oct 2024
281 points (99.0% liked)

Science Memes

11996 readers
2133 users here now

Welcome to c/science_memes @ Mander.xyz!

A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.



Rules

  1. Don't throw mud. Behave like an intellectual and remember the human.
  2. Keep it rooted (on topic).
  3. No spam.
  4. Infographics welcome, get schooled.

This is a science community. We use the Dawkins definition of meme.



Research Committee

Other Mander Communities

Science and Research

Biology and Life Sciences

Physical Sciences

Humanities and Social Sciences

Practical and Applied Sciences

Memes

Miscellaneous

founded 2 years ago
MODERATORS
 
top 7 comments
sorted by: hot top controversial new old
[–] [email protected] 20 points 3 months ago (2 children)

We don't. We keep just doing things and good things keep happening afterwards.

We don't even know if those two facts are linked in any way.

[–] [email protected] 13 points 3 months ago

Nearly irrelevant xkcd

At least in software we know where the linchpins are on some level.

[–] [email protected] 3 points 3 months ago

Descartes said it best. The only thing I can know for sure is that I do, in fact, exist.

[–] [email protected] 11 points 3 months ago (1 children)

Frequentist statistics are really... silly in a way. And this coming from someone who has to teach it. Sure, p is less than 5%, but you sampled 100,000 people-- an effect size of 0.05 would be significant at this rate. "bUt ItS sIgNiFiCaNt"... Oy.

[–] [email protected] 8 points 3 months ago (1 children)

I get very suspicious if a paper samples multiple groups and still uses p. You would use q in that case, and the fact that they didn't suggests that nothing came up positive.

Still, in my opinion it's generally OK if they only use the screen as a starting point and do follow-up experiments afterwards

[–] [email protected] 5 points 3 months ago

Yeah, I used to work in a field with huge samples so significance wasn't really all that useful. I usually just report significant coefficients and try to make clear what changes by model. For instance, if a type of curriculum showed improvements on test scores, you simply say how much and, possibly, illustrate it by saying if a person went from 50th percentile to 55th percentile.

Every field varies, though. I find it crazy how much psychologists I've worked with cared about r-squared. To each their own, I guess.

[–] [email protected] 3 points 3 months ago

The fun part is that we don't