In this article, I will discuss some great tips and tricks to improve the performance of your structured data binary classification model. These tricks are obtained from solutions of some of Kaggle’s top tabular data competitions. Without much lag, let’s begin. These are the five competitions that I have gone through to create this article: Home credit default risk Santander Customer Transaction Prediction VSB Power Line Fault Detection Microsoft Malware Prediction IEEE-CIS Fraud Detection Dealing with larger datasets One issue you might face in any machine learning competition is the size of your data set. If the size of your data is large, that is 3GB + for kaggle kernels and more basic laptops you could find it difficult to load and process with limited resources. Here is the link to some of the articles and kernels that I have found useful in such situations. Faster data loading with pandas. Data compression techniques to . reduce the size of data by 70% Optimize the memory by r educing the size of some attributes. Use open-source libraries such as , it performs parallel computing and saves up memory space. Dask to read and manipulate the data Use . cudf Convert data to format. parquet Converting data to format. feather Reducing memory usage for . optimizing RAM Data exploration Data exploration always helps to better understand the data and gain insights from it. Before starting to develop machine learning models, top competitors always read/do a lot of exploratory data analysis for the data. This helps in feature engineering and cleaning of the data. EDA for microsoft malware detection. Time Series EDA for malware detection. Complete . EDA for home credit loan prediction Complete EDA for Santader prediction. EDA for VSB Power Line Fault Detection. Data preparation After data exploration, the first thing to do is to use those insights to prepare the data. To tackle issues like class imbalance, encoding categorical data, etc. Let’s see the methods used to do it. Methods to t . ackle class imbalance Data augmentation by . Synthetic Minority Oversampling Technique Fast inplace . shuffle for augmentation Finding synthetic samples in the dataset. used in signal processing competitions. Signal denoising Finding . patterns of missing data Methods to handle . missing data An overview of various encoding techniques for categorical data. Building model to predict missing values. Random to create new synthetic training set. shuffling of data Feature engineering Next, you can check the most popular feature and feature engineering techniques used in these top kaggle competitions. The feature engineering part varies from problem to problem depending on the domain. Target for better encoding. encoding cross validation Entity embedding to . handle categories Encoding c yclic features for deep learning. Manual . feature engineering methods Automated feature engineering techniques . using featuretools Top hard crafted features used in . microsoft malware detection Denoising NN for . feature extraction Feature engineering using RAPIDS framework. Things to remember while processing f eatures using LGBM. Lag features and moving averages. for dimensionality reduction. Principal component analysis LDA for . dimensionality reduction Best hand crafted LGBM features for . microsoft malware detection Generating frequency features. Dropping variables with different train and test distribution. for home credit competition. Aggregate time series features used in home credit default risk. Time Series features Scale,Standardize and n . ormalize with sklearn Handcrafted features for Home default risk competition. Handcrafted features used in Santander Transaction Prediction. Feature selection After generating many features from your data, you need to decide which all features to use in your model to get the maximum performance out of your model. This step also includes identifying the impact each feature is having on your model. Let’s see some of the most popular feature selection methods. Six ways to do f . eatures selection using sklearn . Permutation feature importance . Adversarial feature validation Feature selection using null importances. Tree explainer using SHAP. DeepNN explainer using . SHAP Modeling After handcrafting and selecting your features, you should choose the right Machine learning algorithm to make your prediction. These are the collection of some of the most used ML models in structured data classification challenges. Random forest classifier . XGBoost : Gradient boosted decision trees. for distributed and faster training. LightGBM to handle categorical data. CatBoost classifier. Naive bayes model. Gaussian naive bayes used in 3rd place solution of Santander Customer Transaction Prediction LGBM + CNN model n Neural Network. Knowledge distillation i method. Follow the regularized leader Comparison between (goss, gbdt and dart). LGB boosting methods experiment. NN + focal loss Keras NN with timeseries splitter. 5th place for Santander Transaction prediction. NN architecture with code Hyperparameter tuning LGBM methods. hyperparameter tuning Automated methods. model tuning Parameter tuning with . hyperopt for hyperparameter tuning. Bayesian optimization . Gpyopt Hyperparameter Optimisation Evaluation Choosing a suitable validation strategy is very important to avoid huge shake-ups or poor performance of the model in the private test set. The traditional 80:20 split wouldn’t work for many cases. Cross-validation works in most cases over the traditional single train-validation split to estimate the model performance. There are different variations of KFold cross-validation such as group k-fold that should be chosen accordingly. . K-fold cross-validation . Stratified KFold cross-validation Group KFold to check if train and test distributions are similar or not. Adversarial validation validation. Time Series split Extensive time series splitter. Note: There are various metrics that you can use to evaluate the performance of your tabular models. A bunch of useful classification metrics are listed and explained here. Other training tricks for LGBM. GPU acceleration . Use the GPU efficiently . Free keras memory to save runtime and memory. Save and load models Ensemble If you’re in the competing environment one won’t get to the top of the leaderboard without ensembling. Selecting the appropriate ensembling/stacking method is very important to get the maximum performance out of your models. Let’s see some of the popular ensembling techniques used in kaggle competitions: . Weighted average ensemble Stacked generalization ensemble. . Out of folds predictions . Blending with linear regression Use to determine blending weights. optuna . Power average ensemble Power 3.5 blending strategy. . Blending diverse models Different stacking approaches. AUC weight optimization. for low correlation predictions. Geometric mean Weighted . rank average Final thoughts In this article, you saw many popular and effective ways to improve the performance of your tabular data binary classification model. Hopefully, you will find them useful in your projects. This article was originally written by Shahul ES and posted on the Neptune blog . You can find more in-depth articles for machine learning practitioners there.