← Kaggle

📊 Key Metrics

0.947
Public LB (ROC-AUC)
Best submission so far
0.950
5-Fold CV (OOF)
LGB + XGB ensemble
~800
Feature Columns
After feature engineering v1
200
Optuna Trials
TPE sampler per model

🤖 Model Performance

Model3-Fold CV5-Fold OOFHardware
LightGBM (Optuna)0.9500.950GPU (CUDA)
XGBoost (Optuna)0.9500.949GPU (CUDA)
CatBoost (Optuna)0.947GPU
Ensemble (avg)0.9510.950

🔬 Approach

1

Feature Engineering v1

Target encoding on categoricals (Race, Driver, Team), lag features on LapNumber, rolling statistics, interaction terms. ~800 features with StratifiedKFold validation.

2

Model Baseline

LGB + XGB + CatBoost with default params on 5-fold CV. Best single model: XGB 0.950 CV. Ensemble 0.951 CV. Public LB 0.94735.

3

Optuna Hyperparameter Tuning

200-trial TPE search per model. Parameter space: n_estimators 200-1200, lr 0.005-0.15, max_depth 3-12, subsample/colsample 0.5-1.0. Dropped CatBoost (underperforming). Currently running.

4

Next Steps

Blend with public high-score submissions, pseudo-labeling, seed averaging, final 5-fold OOF ensemble with BayesianRidge stacking.

📤 Recent Submissions

Optuna 200t LGB+XGB (running)
V1 3-model ensemble + seed avg 0.94735
V1 LGB+XGB ensemble (5-fold OOF) 0.94696
V1 XGB default 5-fold 0.936xx