10 Lecture 01 - 2019

10.0.1 Popper

  • generate meaningful (not null) hypotheses and predictions and falsify those

10.0.2 Approach

A framework for developing + using statistical golems

  1. Bayesian data analysis
    • uses probability to describe uncertainty
    • “count all the ways data can happen, according to assumptions and the assumptions with more ways consistent with the data are more plausible”
  2. Multilevel models
    • Models within models
    • Avoids averaging
  3. Model comparison
    • Compare meaningful (not null) models
    • Caution: over fitting

10.1 Hypotheses - Process Models - Statistical Models

Any statistical model M can correspond to multiple process models

Any hypothesis H may correspond to multiple process models

Any statistical model may correspond to multiple hypothesis

10.2 Small world / large world

Small world: models have assumptions, Bayesian models fit optimally

Large world: real world, no guarantee of optimality

10.3 Example: four marbles

Setup

4 marbles, either black or white, with replacement

Possibilities (5) therefore: WWWW, BWWW, BBWW, BBBW, BBBB

Observation: BWB

Calculate

Given 3 observations, there are 4 choices, for a total of 64 possibilities

Given we observed both a white and a black marble, possibilities WWWW and BBBB are not valid

At each branch, there are 3 possibilities it can be white and 1 possibility it can be black

Bayesian is additive, at each branch just sum the possibility

BWWW: 3 = 1 * 3 * 1

BBWW: 8 = 2 * 2 * 2

BBBW: 9 = 3 * 1 * 3

Using new information

New information is directly integrated into the old information, therefore just multiply it through

So if we take another measure of B, multiply the property through

BWWW: 3 * 1 = 3

BBWW: 8 * 2 = 16

BBBW: 9 * 3 = 27

Using other information

Factory says B measures are rare, but minimum of 1 per bag

           Factory info

WWWW 0 since we observed a B

BWWW: 3

BBWW: 2

BBBW: 1

BBBB 0 since we observed a W

Multiply it through

BWWW: 3 * 3 = 9

BBWW: 16 * 2 = 32

BBBW: 1 * 27 = 27

Counts get huge - therefore we normalize them giving us probabilities (0-1)

Probability theory is just normalized counting

10.4 Building a model

  1. Design the model
  2. Condition on the data
  3. Evaluate, critique the model
  4. (Restart)