Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Conditional Inference for New Data #15

Open
melondonkey opened this issue May 6, 2021 · 1 comment
Open

Conditional Inference for New Data #15

melondonkey opened this issue May 6, 2021 · 1 comment
Labels
enhancement New feature or request

Comments

@melondonkey
Copy link

Is there a way to use the trained models to do conditional inference on new observations and also get the underlying probabilities rather than sampled data sets? For example, I train on a binary matrix of diagnoses and then as a new patient comes in, I can input their known conditions and get the probability they have the other conditions?

The ability to do that in combination with the TF API would make this a very powerful "auto-complete" model.

@tsrobinson tsrobinson added the enhancement New feature or request label May 11, 2021
@tsrobinson
Copy link
Collaborator

tsrobinson commented May 11, 2021

Thanks for raising this issue!

It is possible to recover the predicted probabilities (rather than labels) by setting cat_coalesce = FALSE and bin_label = FALSE in the complete() function. Since uncertainty over the predictions is handled via multiply imputing the data, the best strategy would then be to average across M completed datasets in order to get good estimates of the average predicted probabilities.

We are actively looking into adding a new function to predict missing values for data not used in training, which would allow you to achieve the proposed workflow above.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants