You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When feature fraction is set to a small value (e.g. 0.6), the last feature in the data set never gets selected.
I was expecting that the feature subset is re-evaluated for each iteration.
In general, a feature fraction that is not close to 1 has some feature_importances==0 for my data-set, but not necessarily in the last position. I suspect there is a bug, either in the description of how this parameter works or the implementation
Reproducible example
I am creating three random features, only the last one is used in the response but it will not get picked up unless
the feature fraction gets increased.
this is visible from feature importance = 0 and from the response plot.
Please increase feature-fraction or change the order of the predictors (e.g. all_preds = ["z", "x", "y"]) to see how the model suddenly learns the function.
Hey @draphi ! Thanks a lot for posting this issue with detailed reproducible example!
I can confirm that one feature is unused in 3.2.1 version. But I think this issue has been fixed in master via #4450. Also linking #4371 as the same issue.
Here is what I get with nightly build of LightGBM:
This issue has been automatically closed because it has been awaiting a response for too long. When you have time to to work with the maintainers to resolve this issue, please post a new comment and it will be re-opened. If the issue has been locked for editing by the time you return to it, please open a new issue and reference this one. Thank you for taking the time to improve LightGBM!
This issue has been automatically locked since there has not been any recent activity since it was closed. To start a new related discussion, open a new issue at https://github.com/microsoft/LightGBM/issues including a reference to this.
Description
When feature fraction is set to a small value (e.g. 0.6), the last feature in the data set never gets selected.
I was expecting that the feature subset is re-evaluated for each iteration.
In general, a feature fraction that is not close to 1 has some feature_importances==0 for my data-set, but not necessarily in the last position. I suspect there is a bug, either in the description of how this parameter works or the implementation
Reproducible example
I am creating three random features, only the last one is used in the response but it will not get picked up unless
the feature fraction gets increased.
this is visible from feature importance = 0 and from the response plot.
Please increase feature-fraction or change the order of the predictors (e.g. all_preds = ["z", "x", "y"]) to see how the model suddenly learns the function.
Environment info
Release version:
pip install lightgbm==3.2.1
Additional Comments
The text was updated successfully, but these errors were encountered: