forked fromqiyiping/gbdt
- Notifications
You must be signed in to change notification settings - Fork0
mtxrym/gbdt
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
- Download the code:
git clone https://github.com/qiyiping/gbdt.git
- Run
make
to compile (boost is required) - Run the demo script:
./demo.sh
[InitalGuess] Label Weight Index0:Value0 Index1:Value1 ..
Each line contains an instance and is ended by a ‘\n’ character. Inital guess is optional. For two-class classification, Label is -1 or 1. For regression, Label is the target value, which can be any real number. Feature Index starts from 0. Feature Value can be any real number.
classConfigure {public:size_tnumber_of_feature;// number of featuressize_tmax_depth;// max depth for each treesize_titerations;// number of trees in gbdtdoubleshrinkage;// shrinkage parameterdoublefeature_sample_ratio;// portion of features to be spliteddoubledata_sample_ratio;// portion of data to be fitted in each iterationsize_tmin_leaf_size;// min number of nodes in leafLossloss;// loss typebooldebug;// show debug info?double*feature_costs;// mannually set feature costs in order to tune the modelboolenable_feature_tunning;// when set true, `feature_costs' is used to tune the modelboolenable_initial_guess;...};
- Friedman, J. H. “Greedy Function Approximation: A Gradient Boosting Machine.” (February 1999)
- Friedman, J. H. “Stochastic Gradient Boosting.” (March 1999)
- Jerry Ye, et al. (2009). Stochastic gradient boosted distributed decision trees. (Distributed implementation)
About
fork gbdt implemention
Resources
License
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
No releases published
Packages0
No packages published
Languages
- C++89.9%
- Python6.6%
- Makefile2.7%
- Shell0.8%