Skip to content

Commit a75da9d

Browse files
committed
new file: packages/orf/orf.1.0.1/opam
compiles w/ ocaml-5 and many other versions probably
1 parent 25c8373 commit a75da9d

File tree

1 file changed

+63
-0
lines changed
  • packages/orf/orf.1.0.1

1 file changed

+63
-0
lines changed

packages/orf/orf.1.0.1/opam

Lines changed: 63 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,63 @@
1+
opam-version: "2.0"
2+
authors: "Francois Berenger"
3+
maintainer: "[email protected]"
4+
homepage: "https://github.com/UnixJunkie/orf"
5+
bug-reports: "https://github.com/UnixJunkie/orf/issues"
6+
dev-repo: "git+https://github.com/UnixJunkie/orf.git"
7+
license: "LGPL-2.1-or-later WITH OCaml-LGPL-linking-exception"
8+
build: ["dune" "build" "-p" name "-j" jobs]
9+
depends: [
10+
"batteries" {>= "3.2.0"}
11+
"cpm" {>= "6.0.0"}
12+
"dolog" {>= "4.0.0"}
13+
"dune" {>= "2.8"}
14+
"minicli"
15+
"ocaml"
16+
"parany" {>= "11.0.0"}
17+
"line_oriented"
18+
]
19+
depopts: [
20+
"conf-gnuplot"
21+
]
22+
synopsis: "OCaml Random Forests"
23+
description:"""
24+
Random Forests (RFs) can do classification or regression modeling.
25+
26+
Random Forests are one of the workhorse of modern machine
27+
learning. Especially, they cannot over-fit to the training set, are
28+
fast to train, predict fast, parallelize well and give you a reasonable
29+
model even without optimizing the model's default hyper-parameters. In
30+
other words, it is hard to shoot yourself in the foot while training
31+
or exploiting a Random Forests model. In comparison, with deep neural
32+
networks it is very easy to shoot yourself in the foot.
33+
34+
Using out of bag (OOB) samples, you can even get an idea of a RFs
35+
performance, without the need for a held out (test) data-set.
36+
37+
Their only drawback is that RFs, being an ensemble model, cannot predict
38+
values which are outside of the training set range of values (this is
39+
a serious limitation in case you are trying to optimize or minimize
40+
something in order to discover outliers, compared to your training
41+
set samples).
42+
43+
For the moment, this implementation only consider a sparse vector
44+
of integers as features. i.e. categorical variables will need to be
45+
one-hot-encoded.
46+
For classification, the dependent variable must be an integer
47+
(encoding a class label).
48+
For regression, the dependent variable must be a float.
49+
50+
Bibliography
51+
============
52+
53+
Breiman, Leo. (1996). Bagging Predictors. Machine learning, 24(2),
54+
123-140.
55+
56+
Breiman, Leo. (2001). Random Forests. Machine learning, 45(1), 5-32.
57+
58+
Geurts, P., Ernst, D., & Wehenkel, L. (2006). Extremely Randomized
59+
Trees. Machine learning, 63(1), 3-42."""
60+
url {
61+
src: "https://github.com/UnixJunkie/orf/archive/refs/tags/v1.0.1.tar.gz"
62+
checksum: "md5=d41d8cd98f00b204e9800998ecf8427e"
63+
}

0 commit comments

Comments
 (0)