Skip to content

[ENH] Gradient Boosting: New widget#5160

Merged
lanzagar merged 11 commits intobiolab:masterfrom
VesnaT:gbtrees
Jan 15, 2021
Merged

[ENH] Gradient Boosting: New widget#5160
lanzagar merged 11 commits intobiolab:masterfrom
VesnaT:gbtrees

Conversation

@VesnaT
Copy link
Copy Markdown
Contributor

@VesnaT VesnaT commented Jan 5, 2021

Issue

Implements #5107
Alternative to #5116

Description of changes
  • wrap sklearn gradient boosting classification and regression learners
  • wrap catboost classification and regression learners
  • wrap xgboost classification and regression learners
  • Gradient Boosting widget
Includes
  • Code changes
  • Tests
  • Documentation

@codecov
Copy link
Copy Markdown

codecov bot commented Jan 5, 2021

Codecov Report

Merging #5160 (bfffd6e) into master (ef789ee) will increase coverage by 0.20%.
The diff coverage is 97.80%.

@@            Coverage Diff             @@
##           master    #5160      +/-   ##
==========================================
+ Coverage   84.77%   84.98%   +0.20%     
==========================================
  Files         286      296      +10     
  Lines       60084    60643     +559     
==========================================
+ Hits        50938    51539     +601     
+ Misses       9146     9104      -42     

@VesnaT VesnaT force-pushed the gbtrees branch 2 times, most recently from 13b88c6 to f2da44e Compare January 6, 2021 08:41
@VesnaT VesnaT changed the title [WIP] Gradient Boosting: Wrapped models, widget [ENH] Gradient Boosting: Wrapped models, widget Jan 7, 2021
@lanzagar
Copy link
Copy Markdown
Contributor

lanzagar commented Jan 8, 2021

Works well for me.
I would suggest:

  • being explicit for the default "Gradient Boosting" method as well and changing to "Gradient Boosting (scikit-learn)" to be more consistent with other entries in the dropdown
  • changing the default num of trees for catboost from 500 to 100 so it is the same as all the other methods - catboost is already the slowest (might need additional investigation)
  • change default value of Replicable training to True. BTW is this really only available from scikit-learn, do xgboost and catboost not allow setting random seeds?

@VesnaT
Copy link
Copy Markdown
Contributor Author

VesnaT commented Jan 11, 2021

  • change default value of Replicable training to True. BTW is this really only available from scikit-learn, do xgboost and catboost not allow setting random seeds?

I think so. Both (xgboost and catboost) have 0 as default. If one passes None, it is changed to 0.

@lanzagar lanzagar changed the title [ENH] Gradient Boosting: Wrapped models, widget [ENH] Gradient Boosting: New widget Jan 15, 2021
@lanzagar lanzagar merged commit 1165816 into biolab:master Jan 15, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants