Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

OCaml Random Forests

License

NotificationsYou must be signed in to change notification settings

UnixJunkie/orf

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Random Forests (RFs) are one of the workhorse of modern machine learning.Especially, they cannot over-fit to the training set, arefast to train, predict fast, parallelize well and give you a reasonable modeleven without optimizing the model's default hyper-parameters.In other words, it is hard to shoot yourself in the foot whiletraining or exploiting a Random Forests model.In comparison, with deep neural networksit is very easy to shoot yourself in the foot.

Using out of bag (OOB) samples, you can even get an ideaof a RFs performance, without the need for a held out(test) dataset.

Their only drawback is that RFs, being an ensemble model,cannot predict values which are outside of the training setrange of values (thisis a serious limitation in case youare trying to optimize or minimize something in order to discoveroutliers, compared to your training set samples).

For the moment, this implementation will only consider a sparse vector ofintegers as features. i.e. categorical variables will need to beone-hot-encoded.

Bibliography

Breiman, Leo. (1996). "Bagging Predictors". Machine learning, 24(2), 123-140.

Breiman, Leo. (2001). "Random Forests". Machine learning, 45(1), 5-32.

Geurts, P., Ernst, D., & Wehenkel, L. (2006). "Extremely Randomized Trees".Machine learning, 63(1), 3-42.


[8]ページ先頭

©2009-2025 Movatter.jp