Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement/Document Causal Workflow #9

Open
6 of 9 tasks
cthoyt opened this issue Oct 24, 2023 · 1 comment
Open
6 of 9 tasks

Implement/Document Causal Workflow #9

cthoyt opened this issue Oct 24, 2023 · 1 comment

Comments

@cthoyt
Copy link
Member

cthoyt commented Oct 24, 2023

What does it mean to be "complete" with a task?

  1. Code has been implemented, a pull request has been made, CI tests all pass, and it has been merged
  2. Documentation on both the main functions used in the high-level workflow as well as a full documentation of what the step does (written as part of the python package documentation).
    • How do you know if the documentation is good? Consider if you started using NumPy but had never seen it before. How much did you rely on the documentation? It needs to give both high-level context (at module level), specific documentation about all of the inputs and outputs (in docstring), and demonstration of specific usage (i.e., with real data, that has real context)
    • All documentation should be written for a person who does not understand the methodology (or care about understanding it) but wants to be oriented so they know when they should use each part, how they should use it (give demonstrations), and how they should interpret the part that comes out

Case Studies

All case studies need to have full, working implementations that have minimal (or ideally, none) data massaging outside of the workflow itself.

Each Needs to include context (how was the network made, what choices were put into it, what is the biology that it's modeling) as well as a guide on how to interpret the results

Each Needs a comparison of the usage of parts of the causal workflow vs not using it. How do you interpret these results? Are they "better", and what is the metric for that? This should answer the question to the reader: why should I bother with this (i.e., Eliater) instead of just directly using y0, ananke, or similar?

@cthoyt
Copy link
Member Author

cthoyt commented Oct 24, 2023

Other issue we talked about: backdoor/frontdoor doesn't get applied by y0 when generating an estimand that can cause different results from ananke's estimation. This might be true, but we need to test this empirically, as it isn't always the case (and we can't lose anything from marking fewer nodes as latent)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant