18864
6691
sklearn.ensemble.forest.RandomForestClassifier
sklearn.RandomForestClassifier
sklearn.ensemble.forest.RandomForestClassifier
66
openml==0.12.2,sklearn==0.18
A random forest classifier.
A random forest is a meta estimator that fits a number of decision tree
classifiers on various sub-samples of the dataset and use averaging to
improve the predictive accuracy and control over-fitting.
The sub-sample size is always the same as the original
input sample size but the samples are drawn with replacement if
`bootstrap=True` (default).
2021-08-13T18:08:58
English
sklearn==0.18
numpy>=1.6.1
scipy>=0.9
bootstrap
boolean
true
Whether bootstrap samples are used when building trees
class_weight
dict
null
"balanced_subsample" or None, optional (default=None)
Weights associated with classes in the form ``{class_label: weight}``
If not given, all classes are supposed to have weight one. For
multi-output problems, a list of dicts can be provided in the same
order as the columns of y
The "balanced" mode uses the values of y to automatically adjust
weights inversely proportional to class frequencies in the input data
as ``n_samples / (n_classes * np.bincount(y))``
The "balanced_subsample" mode is the same as "balanced" except that
weights are computed based on the bootstrap sample for every tree
grown
For multi-output, the weights of each column of y will be multiplied
Note that these weights will be multiplied with sample_weight (passed
through the fit method) if sample_weight is specified.
criterion
string
"gini"
The function to measure the quality of a split. Supported criteria are
"gini" for the Gini impurity and "entropy" for the information gain
Note: this parameter is tree-specific
max_depth
integer or None
null
The maximum depth of the tree. If None, then nodes are expanded until
all leaves are pure or until all leaves contain less than
min_samples_split samples
max_features
int
"auto"
The number of features to consider when looking for the best split:
- If int, then consider `max_features` features at each split
- If float, then `max_features` is a percentage and
`int(max_features * n_features)` features are considered at each
split
- If "auto", then `max_features=sqrt(n_features)`
- If "sqrt", then `max_features=sqrt(n_features)` (same as "auto")
- If "log2", then `max_features=log2(n_features)`
- If None, then `max_features=n_features`
Note: the search for a split does not stop until at least one
valid partition of the node samples is found, even if it requires to
effectively inspect more than ``max_features`` features
max_leaf_nodes
int or None
null
Grow trees with ``max_leaf_nodes`` in best-first fashion
Best nodes are defined as relative reduction in impurity
If None then unlimited number of leaf nodes
min_impurity_split
float
1e-07
Threshold for early stopping in tree growth. A node will split
if its impurity is above the threshold, otherwise it is a leaf
.. versionadded:: 0.18
min_samples_leaf
int
1
The minimum number of samples required to be at a leaf node:
- If int, then consider `min_samples_leaf` as the minimum number
- If float, then `min_samples_leaf` is a percentage and
`ceil(min_samples_leaf * n_samples)` are the minimum
number of samples for each node
.. versionchanged:: 0.18
Added float values for percentages
min_samples_split
int
2
The minimum number of samples required to split an internal node:
- If int, then consider `min_samples_split` as the minimum number
- If float, then `min_samples_split` is a percentage and
`ceil(min_samples_split * n_samples)` are the minimum
number of samples for each split
.. versionchanged:: 0.18
Added float values for percentages
min_weight_fraction_leaf
float
0.0
The minimum weighted fraction of the input samples required to be at a
leaf node
n_estimators
integer
10
The number of trees in the forest
n_jobs
integer
1
The number of jobs to run in parallel for both `fit` and `predict`
If -1, then the number of jobs is set to the number of cores
oob_score
bool
false
Whether to use out-of-bag samples to estimate
the generalization accuracy
random_state
int
null
If int, random_state is the seed used by the random number generator;
If RandomState instance, random_state is the random number generator;
If None, the random number generator is the RandomState instance used
by `np.random`
verbose
int
0
Controls the verbosity of the tree building process
warm_start
bool
false
When set to ``True``, reuse the solution of the previous call to fit
and add more estimators to the ensemble, otherwise, just fit a whole
new forest
openml-python
python
scikit-learn
sklearn
sklearn_0.18