-
Notifications
You must be signed in to change notification settings - Fork 967
Tests for causal prediction #1321
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Tests for causal prediction #1321
Conversation
5df95b8
to
58a2504
Compare
58a2504
to
424755d
Compare
Signed-off-by: maartenvanhooft <[email protected]>
424755d
to
17a97cb
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for adding this PR, @maartenvanhooftds . The tests make sense, but I'm wondering if we can have a stronger test that compares CACM and ERM.
How about the following property: difference in accuracy between a test dataset from the same distribution as train, and the main test dataset? That would be higher for ERM and we can check as a comparison assert. Can you add this for your setup?
results = trainer.test(algorithm, dataloaders=loaders["test_loaders"]) | ||
assert isinstance(results, list) | ||
assert len(results) > 0 | ||
for r in results: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
does the code assume that both ERM and CACM will lead to >0.7 accuracy? What is the exclusive property that we can testing for CACM?
Great feedback, thanks! Will implement it later this week. Edit: Sorry, it has been taking a bit longer, just started a new job. It's still on my mind though. |
Signed-off-by: maartenvanhooft <[email protected]>
2fec8f0
to
f228a4e
Compare
Inspired by Issue 1313.
Changes:
First contribution here, please review critically for any mistakes or inconsistencies!