A lazy bagging classifier.
Everything is done lazily, models are built at prediction time and are not kept in memory. Since the models is thrown away, this allows to highly reduce the memory consumption and allows to build very large ensemble.
Parameters: | base_estimator : object or None, optional (default=None)
n_estimators : int, optional (default=10)
max_samples : int or float, optional (default=1.0)
max_features : int or float, optional (default=1.0)
bootstrap : boolean, optional (default=True)
bootstrap_features : boolean, optional (default=False)
random_state : int, RandomState instance or None, optional (default=None)
verbose : int, optional (default=0)
|
---|
References
[R5] | L. Breiman, “Pasting small votes for classification in large databases and on-line”, Machine Learning, 36(1), 85-103, 1999. |
[R6] | L. Breiman, “Bagging predictors”, Machine Learning, 24(2), 123-140, 1996. |
[R7] | T. Ho, “The random subspace method for constructing decision forests”, Pattern Analysis and Machine Intelligence, 20(8), 832-844, 1998. |
[R8] | G. Louppe and P. Geurts, “Ensembles on Random Patches”, Machine Learning and Knowledge Discovery in Databases, 346-361, 2012. |
Attributes
classes_ | array of shape = [n_classes] | The classes labels. |
n_classes_ | int or list | The number of classes. |
n_features_ | int, | Number of features of the fitted input matrix |
n_outputs_ | int, | Number of outputs of the fitted ouput matrix |
random_seed_ | int, | Seed of the number generator |
Methods
decision_function(X) | Average of the decision functions of the base classifiers. |
fit(X, y[, sample_weight]) | Build a lazy a bagging ensemble of estimators from the training set |
get_params([deep]) | Get parameters for this estimator. |
predict(X) | Predict class for X. |
predict_log_proba(X) | Predict class log-probabilities for X. |
predict_proba(X) | Predict class probabilities for X. |
score(X, y[, sample_weight]) | Returns the mean accuracy on the given test data and labels. |
set_params(**params) | Set the parameters of this estimator. |
Average of the decision functions of the base classifiers.
Parameters: | X : {array-like, sparse matrix} of shape = [n_samples, n_features]
|
---|---|
Returns: | score : array, shape = [n_samples, k] or list of array
|
Build a lazy a bagging ensemble of estimators from the training set
Parameters: | X : {array-like, sparse matrix} of shape = [n_samples, n_features]
y : array-like, shape = [n_samples]
sample_weight : array-like, shape = [n_samples] or None
|
---|---|
Returns: | self : object
|
Get parameters for this estimator.
Parameters: | deep: boolean, optional :
|
---|---|
Returns: | params : mapping of string to any
|
Predict class for X.
The predicted class of an input sample is computed as the class with the highest mean predicted probability. If base estimators do not implement a predict_proba method, then it resorts to voting.
Parameters: | X : {array-like, sparse matrix} of shape = [n_samples, n_features]
|
---|---|
Returns: | y : array of shape = [n_samples, n_outputs]
|
Predict class log-probabilities for X.
The predicted class log-probabilities of an input sample is computed as the log of the mean predicted class probabilities of the trees in the forest.
Parameters: | X : array-like of shape = [n_samples, n_features]
|
---|---|
Returns: | p : array of shape = [n_samples, n_classes], or a list of n_outputs
|
Predict class probabilities for X.
The predicted class probabilities of an input sample is computed as the mean predicted class probabilities of the base estimators in the ensemble. If base estimators do not implement a predict_proba method, then it resorts to voting and the predicted class probabilities of a an input sample represents the proportion of estimators predicting each class.
Parameters: | X : {array-like, sparse matrix} of shape = [n_samples, n_features]
|
---|---|
Returns: | p : array of shape = [n_samples, n_classes] or list of array
|
Returns the mean accuracy on the given test data and labels.
In multi-label classification, this is the subset accuracy which is a harsh metric since you require for each sample that each label set be correctly predicted.
Parameters: | X : array-like, shape = (n_samples, n_features)
y : array-like, shape = (n_samples) or (n_samples, n_outputs)
sample_weight : array-like, shape = [n_samples], optional
|
---|---|
Returns: | score : float
|
Set the parameters of this estimator.
The method works on simple estimators as well as on nested objects (such as pipelines). The former have parameters of the form <component>__<parameter> so that it’s possible to update each component of a nested object.
Returns: | self : |
---|