-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bias in UncertaintyForest performance compared to paper #377
Comments
@EYezerets as Richard is not a contributor he has to comment on this issue to be assigned. |
@rguo123 i can't assign you, so i slacked you ;) |
@EYezerets Has there been enough progress on this to close this issue? |
@levinwil Sorry Will, we haven't really had any new ideas on this recently. Is it impeding progress on the repo? |
@EYezerets One difference I found between the original notebook and the new benchmark functions is that in the |
To put it more specifically, the original UF has the |DP| : |DV| : |DE| ratio as |
My issue is about the fact that the UncertaintyForest benchmarks notebook shows that the UncertaintyForest class from ProgLearn underperforms IRF at d=20, which we did not see in the original paper.
I checked that samples are taken without replacement now in both the deprecated uncertainty-forest repo and in ProgLearn, i.e.
bootstrap = False
in the figure 2 tutorial in the uncertainty-forest repo, andreplace = False
in progressive-learner.py in ProgLearn. Also, I believe that the n_estimators (300), tree_construction_proportion (0.4), and kappa (3) values are the same.Snapshot of documentation error:
From the paper (original Figure 2):
From benchmarks in EYezerets/ProgLearn on the fig2benchmark branch:
Additional context
Sorry, for some reason I'm not able to assign Richard to this issue. Could someone please help me include him in this conversation?
The text was updated successfully, but these errors were encountered: