A Gentle Introduction to Computational Learning Theory

Last Updated on September 7, 2020 Computational learning theory, or statistical learning theory, refers to mathematical frameworks for quantifying learning tasks and algorithms. These are sub-fields of machine learning that a machine learning practitioner does not need to know in great depth in order to achieve good results on a wide range of problems. Nevertheless, it is a sub-field where having a high-level understanding of some of the more prominent methods may provide insight into the broader task of learning […]

Read more

Plot a Decision Surface for Machine Learning Algorithms in Python

Last Updated on August 26, 2020 Classification algorithms learn how to assign class labels to examples, although their decisions can appear opaque. A popular diagnostic for understanding the decisions made by a classification algorithm is the decision surface. This is a plot that shows how a fit machine learning algorithm predicts a coarse grid across the input feature space. A decision surface plot is a powerful tool for understanding how a given model “sees” the prediction task and how it […]

Read more

Why Do I Get Different Results Each Time in Machine Learning?

Last Updated on August 27, 2020 Are you getting different results for your machine learning algorithm? Perhaps your results differ from a tutorial and you want to understand why. Perhaps your model is making different predictions each time it is trained, even when it is trained on the same data set each time. This is to be expected and might even be a feature of the algorithm, not a bug. In this tutorial, you will discover why you can expect […]

Read more

How to Calculate the Bias-Variance Trade-off with Python

Last Updated on August 26, 2020 The performance of a machine learning model can be characterized in terms of the bias and the variance of the model. A model with high bias makes strong assumptions about the form of the unknown underlying function that maps inputs to outputs in the dataset, such as linear regression. A model with high variance is highly dependent upon the specifics of the training dataset, such as unpruned decision trees. We desire models with low […]

Read more

Hypothesis Test for Comparing Machine Learning Algorithms

Last Updated on September 1, 2020 Machine learning models are chosen based on their mean performance, often calculated using k-fold cross-validation. The algorithm with the best mean performance is expected to be better than those algorithms with worse mean performance. But what if the difference in the mean performance is caused by a statistical fluke? The solution is to use a statistical hypothesis test to evaluate whether the difference in the mean performance between any two algorithms is real or […]

Read more

How to Set Axis for Rows and Columns in NumPy

NumPy arrays provide a fast and efficient way to store and manipulate data in Python. They are particularly useful for representing data as vectors and matrices in machine learning. Data in NumPy arrays can be accessed directly via column and row indexes, and this is reasonably straightforward. Nevertheless, sometimes we must perform operations on arrays of data such as sum or mean of values by row or column and this requires the axis of the operation to be specified. Unfortunately, […]

Read more

Time Series Forecasting With Prophet in Python

Time series forecasting can be challenging as there are many different methods you could use and many different hyperparameters for each method. The Prophet library is an open-source library designed for making forecasts for univariate time series datasets. It is easy to use and designed to automatically find a good set of hyperparameters for the model in an effort to make skillful forecasts for data with trends and seasonal structure by default. In this tutorial, you will discover how to […]

Read more

Deep Learning Models for Multi-Output Regression

Last Updated on August 28, 2020 Multi-output regression involves predicting two or more numerical variables. Unlike normal regression where a single value is predicted for each sample, multi-output regression requires specialized machine learning algorithms that support outputting multiple variables for each prediction. Deep learning neural networks are an example of an algorithm that natively supports multi-output regression problems. Neural network models for multi-output regression tasks can be easily defined and evaluated using the Keras deep learning library. In this tutorial, […]

Read more

Multi-Label Classification with Deep Learning

Last Updated on August 31, 2020 Multi-label classification involves predicting zero or more class labels. Unlike normal classification tasks where class labels are mutually exclusive, multi-label classification requires specialized machine learning algorithms that support predicting multiple mutually non-exclusive classes or “labels.” Deep learning neural networks are an example of an algorithm that natively supports multi-label classification problems. Neural network models for multi-label classification tasks can be easily defined and evaluated using the Keras deep learning library. In this tutorial, you […]

Read more

How to Use AutoKeras for Classification and Regression

Last Updated on September 6, 2020 AutoML refers to techniques for automatically discovering the best-performing model for a given dataset. When applied to neural networks, this involves both discovering the model architecture and the hyperparameters used to train the model, generally referred to as neural architecture search. AutoKeras is an open-source library for performing AutoML for deep learning models. The search is performed using so-called Keras models via the TensorFlow tf.keras API. It provides a simple and effective approach for […]

Read more
1 842 843 844 845 846 896