XGBoosting Home | About | Contact | Examples

Configure XGBoost "reg_alpha" Parameter

The reg_alpha parameter in XGBoost is an alias for the alpha parameter, which controls the L1 regularization term on weights. By adjusting reg_alpha, you can influence the model’s complexity and sparsity.

from xgboost import XGBClassifier
from sklearn.datasets import make_classification
from sklearn.model_selection import train_test_split

# Generate synthetic data
X, y = make_classification(n_samples=1000, n_features=20, n_informative=2, n_redundant=10, random_state=42)

# Split the dataset into training and test sets
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)

# Initialize the XGBoost classifier with an alpha value
model = XGBClassifier(reg_alpha=0.1, eval_metric='logloss')

# Fit the model
model.fit(X_train, y_train)

# Make predictions
predictions = model.predict(X_test)

As discussed in the example on configuring the alpha parameter, reg_alpha determines the strength of the L1 regularization term on the weights in the XGBoost model. It is a regularization parameter that can help prevent overfitting and promote sparsity by encouraging the model to use fewer features. reg_alpha accepts non-negative values, and the default value in XGBoost is 0, which means no L1 regularization is applied.

To recap, the key points when configuring the reg_alpha parameter are:

For practical guidance on choosing the right reg_alpha value, refer to the example on configuring the alpha parameter.

See Also