Articles About Machine Learning

How to Grid Search Data Preparation Techniques

Last Updated on August 17, 2020 Machine learning predictive modeling performance is only as good as your data, and your data is only as good as the way you prepare it for modeling. The most common approach to data preparation is to study a dataset and review the expectations of a machine learning algorithms, then carefully choose the most appropriate data preparation techniques to transform the raw data to best meet the expectations of the algorithm. This is slow, expensive, […]

Read more

How to Create Custom Data Transforms for Scikit-Learn

Last Updated on July 19, 2020 The scikit-learn Python library for machine learning offers a suite of data transforms for changing the scale and distribution of input data, as well as removing input features (columns). There are many simple data cleaning operations, such as removing outliers and removing columns with few observations, that are often performed manually to the data, requiring custom code. The scikit-learn library provides a way to wrap these custom data transforms in a standard way so […]

Read more

Add Binary Flags for Missing Values for Machine Learning

Last Updated on August 17, 2020 Missing values can cause problems when modeling classification and regression prediction problems with machine learning algorithms. A common approach is to replace missing values with a calculated statistic, such as the mean of the column. This allows the dataset to be modeled as per normal but gives no indication to the model that the row original contained missing values. One approach to address this issue is to include additional binary flag input features that […]

Read more

How to Selectively Scale Numerical Input Variables for Machine Learning

Last Updated on August 17, 2020 Many machine learning models perform better when input variables are carefully transformed or scaled prior to modeling. It is convenient, and therefore common, to apply the same data transforms, such as standardization and normalization, equally to all input variables. This can achieve good results on many problems. Nevertheless, better results may be achieved by carefully selecting which data transform to apply to each input variable prior to modeling. In this tutorial, you will discover […]

Read more

Train-Test Split for Evaluating Machine Learning Algorithms

Last Updated on August 26, 2020 The train-test split procedure is used to estimate the performance of machine learning algorithms when they are used to make predictions on data not used to train the model. It is a fast and easy procedure to perform, the results of which allow you to compare the performance of machine learning algorithms for your predictive modeling problem. Although simple to use and interpret, there are times when the procedure should not be used, such […]

Read more

LOOCV for Evaluating Machine Learning Algorithms

Last Updated on August 26, 2020 The Leave-One-Out Cross-Validation, or LOOCV, procedure is used to estimate the performance of machine learning algorithms when they are used to make predictions on data not used to train the model. It is a computationally expensive procedure to perform, although it results in a reliable and unbiased estimate of model performance. Although simple to use and no configuration to specify, there are times when the procedure should not be used, such as when you […]

Read more

Nested Cross-Validation for Machine Learning with Python

Last Updated on August 28, 2020 The k-fold cross-validation procedure is used to estimate the performance of machine learning models when making predictions on data not used during training. This procedure can be used both when optimizing the hyperparameters of a model on a dataset, and when comparing and selecting a model for the dataset. When the same cross-validation procedure and dataset are used to both tune and select a model, it is likely to lead to an optimistically biased […]

Read more

How to Configure k-Fold Cross-Validation

Last Updated on August 26, 2020 The k-fold cross-validation procedure is a standard method for estimating the performance of a machine learning algorithm on a dataset. A common value for k is 10, although how do we know that this configuration is appropriate for our dataset and our algorithms? One approach is to explore the effect of different k values on the estimate of model performance and compare this to an ideal test condition. This can help to choose an […]

Read more

Repeated k-Fold Cross-Validation for Model Evaluation in Python

Last Updated on August 26, 2020 The k-fold cross-validation procedure is a standard method for estimating the performance of a machine learning algorithm or configuration on a dataset. A single run of the k-fold cross-validation procedure may result in a noisy estimate of model performance. Different splits of the data may result in very different results. Repeated k-fold cross-validation provides a way to improve the estimated performance of a machine learning model. This involves simply repeating the cross-validation procedure multiple […]

Read more

How to Use XGBoost for Time Series Forecasting

Last Updated on August 27, 2020 XGBoost is an efficient implementation of gradient boosting for classification and regression problems. It is both fast and efficient, performing well, if not the best, on a wide range of predictive modeling tasks and is a favorite among data science competition winners, such as those on Kaggle. XGBoost can also be used for time series forecasting, although it requires that the time series dataset be transformed into a supervised learning problem first. It also […]

Read more
1 209 210 211 212 213 222