XGBoosting Home | About | Contact | Examples

XGBoost Configure "scale_pos_weight" Parameter

When working with imbalanced datasets, where the number of instances in each class is significantly different, it’s crucial to adjust the scale_pos_weight parameter in XGBoost to improve model performance.

This parameter controls the balance between positive and negative weights during training, allowing the model to give more importance to the minority class.

The scale_pos_weight parameter should be set to the ratio of negative instances to positive instances in the dataset.

This tells the model how many negative instances (labeled as “0”) there are for each positive instance (labeled as “1”). By setting this parameter correctly, the model can effectively learn from the imbalanced data and make more accurate predictions.

Here’s an example of how to set the scale_pos_weight parameter in XGBoost:

from sklearn.datasets import make_classification
from sklearn.model_selection import train_test_split
from xgboost import XGBClassifier
from sklearn.metrics import confusion_matrix, classification_report

# Generate an imbalanced synthetic dataset
X, y = make_classification(n_samples=1000, n_classes=2, weights=[0.9, 0.1], random_state=42)

# Split data into train and test sets
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)

# Calculate the scale_pos_weight
scale_pos_weight = len(y_train[y_train == 0]) / len(y_train[y_train == 1])

# Initialize an XGBClassifier with the computed scale_pos_weight
model = XGBClassifier(
    n_estimators=100,
    objective='binary:logistic',
    scale_pos_weight=scale_pos_weight,
    random_state=42
)

# Train the model on the imbalanced dataset
model.fit(X_train, y_train)

# Make predictions and evaluate the model's performance
predictions = model.predict(X_test)
print("Confusion Matrix:")
print(confusion_matrix(y_test, predictions))
print("\nClassification Report:")
print(classification_report(y_test, predictions))

In this example, we generate a synthetic imbalanced dataset using make_classification from scikit-learn, with a 9:1 ratio of negative to positive instances. We then calculate the scale_pos_weight as the ratio of negative instances to positive instances in the training set.

Next, we initialize an XGBClassifier with the computed scale_pos_weight and train the model on the imbalanced dataset. Finally, we make predictions on the test set and evaluate the model’s performance using a confusion matrix and classification report.

By setting the scale_pos_weight parameter correctly, the model can effectively handle the class imbalance and improve its performance on the minority class. The confusion matrix and classification report provide insights into the model’s ability to correctly classify instances from both classes, helping you assess its effectiveness on imbalanced data.



See Also