Base learner: Extreme gradient boosting (XGBoost)
Source:R/pipeline_base_functions.R
fit_base_xgb.Rd
XGBoost model is fitted at the defined rate (r_subsample
) of
the input dataset by grid search.
With proper settings, users can utilize graphics
processing units (GPU) to speed up the training process.
Arguments
- dt_imputed
The input data table to be used for fitting.
- folds
pre-generated rset object with minimal number of columns. If NULL,
vfold
should be numeric to be used in rsample::vfold_cv.- tune_mode
character(1). Hyperparameter tuning mode. Default is "grid", "bayes" is acceptable.
- tune_bayes_iter
integer(1). The number of iterations for Bayesian optimization. Default is 50. Only used when
tune_mode = "bayes"
.- learn_rate
The learning rate for the model. For branching purpose. Default is 0.1.
- yvar
The target variable.
- xvar
The predictor variables.
- vfold
The number of folds for cross-validation.
- device
The device to be used for training. Default is "cuda:0". Make sure that your system is equipped with CUDA-enabled graphical processing units.
- trim_resamples
logical(1). Default is TRUE, which replaces the actual data.frames in splits column of
tune_results
object with NA.- return_best
logical(1). If TRUE, the best tuned model is returned.
- ...
Additional arguments to be passed.