Skip to content

Commit

Permalink
updated readme.
Browse files Browse the repository at this point in the history
  • Loading branch information
awanafiaz committed Oct 15, 2024
1 parent f3af01a commit 780a4ac
Show file tree
Hide file tree
Showing 2 changed files with 10 additions and 10 deletions.
10 changes: 5 additions & 5 deletions README.Rmd
Original file line number Diff line number Diff line change
Expand Up @@ -28,9 +28,9 @@ knitr::opts_chunk$set(collapse = TRUE, comment = "#>",

Using predictions from pre-trained algorithms as outcomes in downstream statistical analyses can lead to biased estimates and misleading conclusions. The statistical challenges encountered when drawing inference on predicted data (IPD) include:

1. Understanding the relationship between predicted outcomes and their true, unobserved counterparts
2. Quantifying the robustness of the AI/ML models to resampling or uncertainty about the training data
3. Appropriately propagating both bias and uncertainty from predictions into downstream inferential tasks
1. Understanding the relationship between predicted outcomes and their true, unobserved counterparts.
2. Quantifying the robustness of the AI/ML models to resampling or uncertainty about the training data.
3. Appropriately propagating both bias and uncertainty from predictions into downstream inferential tasks.

Several works have proposed methods for IPD, including post-prediction inference (PostPI) by [Wang et al., 2020](https://www.pnas.org/doi/suppl/10.1073/pnas.2001238117), prediction-powered inference (PPI) and PPI++ by [Angelopoulos et al., 2023a](https://www.science.org/doi/10.1126/science.adi6000) and [Angelopoulos et al., 2023b](https://arxiv.org/abs/2311.01453), and post-prediction adaptive inference (PSPA) by [Miao et al., 2023](https://arxiv.org/abs/2311.14220). Each method was developed to perform inference on a quantity such as the outcome mean or quantile, or a regression coefficient, when we have:

Expand Down Expand Up @@ -192,8 +192,8 @@ fig1

We can see that:

- The predicted outcomes are more correlated with the covariate than the true outcomes (plot A)
- The predicted outcomes are not perfect substitutes for the true outcomes (plot B)
- The predicted outcomes are more correlated with the covariate than the true outcomes (plot A).
- The predicted outcomes are not perfect substitutes for the true outcomes (plot B).

### Model Fitting

Expand Down
10 changes: 5 additions & 5 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -30,11 +30,11 @@ conclusions. The statistical challenges encountered when drawing
inference on predicted data (IPD) include:

1. Understanding the relationship between predicted outcomes and their
true, unobserved counterparts
true, unobserved counterparts.
2. Quantifying the robustness of the AI/ML models to resampling or
uncertainty about the training data
uncertainty about the training data.
3. Appropriately propagating both bias and uncertainty from predictions
into downstream inferential tasks
into downstream inferential tasks.

Several works have proposed methods for IPD, including post-prediction
inference (PostPI) by [Wang et al.,
Expand Down Expand Up @@ -160,9 +160,9 @@ relationships between these variables:
We can see that:

- The predicted outcomes are more correlated with the covariate than the
true outcomes (plot A)
true outcomes (plot A).
- The predicted outcomes are not perfect substitutes for the true
outcomes (plot B)
outcomes (plot B).

### Model Fitting

Expand Down

0 comments on commit 780a4ac

Please sign in to comment.